Browse Source
Merge pull request #441 from hbrunn/8.0-base_partner_merge-cron
Merge pull request #441 from hbrunn/8.0-base_partner_merge-cron
[IMP] backport 9.0 mechanism8.0
Ronald Portier
6 years ago
committed by
GitHub
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
14 changed files with 244 additions and 1060 deletions
-
47base_partner_merge/README.rst
-
4base_partner_merge/__init__.py
-
15base_partner_merge/__openerp__.py
-
915base_partner_merge/base_partner_merge.py
-
29base_partner_merge/hooks.py
-
5base_partner_merge/models/__init__.py
-
16base_partner_merge/models/base_partner_merge.py
-
30base_partner_merge/models/base_partner_merge_automatic_wizard.py
-
21base_partner_merge/models/res_partner.py
-
3base_partner_merge/tests/__init__.py
-
79base_partner_merge/tests/test_merge.py
-
137base_partner_merge/validate_email.py
-
1base_partner_merge/views/base_partner_merge.xml
-
2oca_dependencies.txt
@ -1,2 +1,4 @@ |
|||||
# -*- coding: utf-8 -*- |
# -*- coding: utf-8 -*- |
||||
from . import base_partner_merge # NOQA |
|
||||
|
|
||||
|
from . import models |
||||
|
from .hooks import post_load_hook |
@ -1,17 +1,22 @@ |
|||||
# -*- coding: utf-8 -*- |
# -*- coding: utf-8 -*- |
||||
{ |
{ |
||||
'name': 'Base Partner Merge', |
|
||||
'author': "OpenERP S.A.,Odoo Community Association (OCA)", |
|
||||
|
'name': "Deduplicate Contacts (OCA)", |
||||
|
'author': "Camptocamp,Sunflower IT,Odoo Community Association (OCA)", |
||||
'category': 'Generic Modules/Base', |
'category': 'Generic Modules/Base', |
||||
'version': '8.0.0.1.0', |
|
||||
|
'version': '8.0.1.0.0', |
||||
'license': 'AGPL-3', |
'license': 'AGPL-3', |
||||
'depends': [ |
'depends': [ |
||||
'base', |
'base', |
||||
'mail' |
|
||||
|
'mail', |
||||
|
'base_manifest_extension', |
||||
|
], |
||||
|
'depends_if_installed': [ |
||||
|
'crm', |
||||
], |
], |
||||
'data': [ |
'data': [ |
||||
'security/ir.model.access.csv', |
'security/ir.model.access.csv', |
||||
'base_partner_merge_view.xml', |
|
||||
|
'views/base_partner_merge.xml', |
||||
], |
], |
||||
'installable': True, |
'installable': True, |
||||
|
'post_load': 'post_load_hook', |
||||
} |
} |
@ -1,915 +0,0 @@ |
|||||
# -*- coding: utf-8 -*- |
|
||||
|
|
||||
from __future__ import absolute_import |
|
||||
from email.utils import parseaddr |
|
||||
import functools |
|
||||
import htmlentitydefs |
|
||||
import itertools |
|
||||
import logging |
|
||||
import operator |
|
||||
import re |
|
||||
from ast import literal_eval |
|
||||
from openerp.tools import mute_logger |
|
||||
|
|
||||
# Validation Library https://pypi.python.org/pypi/validate_email/1.1 |
|
||||
from .validate_email import validate_email |
|
||||
|
|
||||
import openerp |
|
||||
import openerp.osv.fields as fields |
|
||||
from openerp.osv.orm import TransientModel, browse_record |
|
||||
from openerp.exceptions import except_orm |
|
||||
from openerp.tools.translate import _ |
|
||||
|
|
||||
pattern = re.compile(r"&(\w+?);") |
|
||||
|
|
||||
_logger = logging.getLogger('base.partner.merge') |
|
||||
|
|
||||
|
|
||||
# http://www.php2python.com/wiki/function.html-entity-decode/ |
|
||||
def html_entity_decode_char(m, defs=None): |
|
||||
if defs is None: |
|
||||
defs = htmlentitydefs.entitydefs |
|
||||
try: |
|
||||
return defs[m.group(1)] |
|
||||
except KeyError: |
|
||||
return m.group(0) |
|
||||
|
|
||||
|
|
||||
def html_entity_decode(string): |
|
||||
return pattern.sub(html_entity_decode_char, string) |
|
||||
|
|
||||
|
|
||||
def sanitize_email(partner_email): |
|
||||
assert isinstance(partner_email, basestring) and partner_email |
|
||||
|
|
||||
result = re.subn(r';|/|:', ',', |
|
||||
html_entity_decode(partner_email or ''))[0].split(',') |
|
||||
|
|
||||
emails = [parseaddr(email)[1] |
|
||||
for item in result |
|
||||
for email in item.split()] |
|
||||
|
|
||||
return [email.lower() |
|
||||
for email in emails |
|
||||
if validate_email(email)] |
|
||||
|
|
||||
|
|
||||
def is_integer_list(ids): |
|
||||
return all(isinstance(i, (int, long)) for i in ids) |
|
||||
|
|
||||
|
|
||||
class MergePartnerLine(TransientModel): |
|
||||
_name = 'base.partner.merge.line' |
|
||||
|
|
||||
_columns = { |
|
||||
'wizard_id': fields.many2one('base.partner.merge.automatic.wizard', |
|
||||
'Wizard'), |
|
||||
'min_id': fields.integer('MinID'), |
|
||||
'aggr_ids': fields.char('Ids', required=True), |
|
||||
} |
|
||||
|
|
||||
_order = 'min_id asc' |
|
||||
|
|
||||
|
|
||||
class MergePartnerAutomatic(TransientModel): |
|
||||
""" |
|
||||
The idea behind this wizard is to create a list of potential partners to |
|
||||
merge. We use two objects, the first one is the wizard for the end-user. |
|
||||
And the second will contain the partner list to merge. |
|
||||
|
|
||||
""" |
|
||||
_name = 'base.partner.merge.automatic.wizard' |
|
||||
|
|
||||
_columns = { |
|
||||
# Group by |
|
||||
'group_by_email': fields.boolean('Email'), |
|
||||
'group_by_name': fields.boolean('Name'), |
|
||||
'group_by_is_company': fields.boolean('Is Company'), |
|
||||
'group_by_vat': fields.boolean('VAT'), |
|
||||
'group_by_parent_id': fields.boolean('Parent Company'), |
|
||||
|
|
||||
'state': fields.selection([('option', 'Option'), |
|
||||
('selection', 'Selection'), |
|
||||
('finished', 'Finished')], |
|
||||
'State', |
|
||||
readonly=True, |
|
||||
required=True), |
|
||||
'number_group': fields.integer("Group of Contacts", readonly=True), |
|
||||
'current_line_id': fields.many2one('base.partner.merge.line', |
|
||||
'Current Line'), |
|
||||
'line_ids': fields.one2many('base.partner.merge.line', |
|
||||
'wizard_id', 'Lines'), |
|
||||
'partner_ids': fields.many2many('res.partner', string='Contacts'), |
|
||||
'dst_partner_id': fields.many2one('res.partner', |
|
||||
string='Destination Contact'), |
|
||||
|
|
||||
'exclude_contact': fields.boolean('A user associated to the contact'), |
|
||||
'exclude_journal_item': fields.boolean('Journal Items associated' |
|
||||
' to the contact'), |
|
||||
'maximum_group': fields.integer("Maximum of Group of Contacts"), |
|
||||
} |
|
||||
|
|
||||
def default_get(self, cr, uid, fields, context=None): |
|
||||
if context is None: |
|
||||
context = {} |
|
||||
res = super(MergePartnerAutomatic, self |
|
||||
).default_get(cr, uid, fields, context) |
|
||||
if (context.get('active_model') == 'res.partner' and |
|
||||
context.get('active_ids')): |
|
||||
partner_ids = context['active_ids'] |
|
||||
res['state'] = 'selection' |
|
||||
res['partner_ids'] = partner_ids |
|
||||
res['dst_partner_id'] = self._get_ordered_partner(cr, uid, |
|
||||
partner_ids, |
|
||||
context=context |
|
||||
)[-1].id |
|
||||
return res |
|
||||
|
|
||||
_defaults = { |
|
||||
'state': 'option' |
|
||||
} |
|
||||
|
|
||||
def get_fk_on(self, cr, table): |
|
||||
q = """ SELECT cl1.relname as table, |
|
||||
att1.attname as column |
|
||||
FROM pg_constraint as con, pg_class as cl1, pg_class as cl2, |
|
||||
pg_attribute as att1, pg_attribute as att2 |
|
||||
WHERE con.conrelid = cl1.oid |
|
||||
AND con.confrelid = cl2.oid |
|
||||
AND array_lower(con.conkey, 1) = 1 |
|
||||
AND con.conkey[1] = att1.attnum |
|
||||
AND att1.attrelid = cl1.oid |
|
||||
AND cl2.relname = %s |
|
||||
AND att2.attname = 'id' |
|
||||
AND array_lower(con.confkey, 1) = 1 |
|
||||
AND con.confkey[1] = att2.attnum |
|
||||
AND att2.attrelid = cl2.oid |
|
||||
AND con.contype = 'f' |
|
||||
""" |
|
||||
return cr.execute(q, (table,)) |
|
||||
|
|
||||
def _update_foreign_keys(self, cr, uid, src_partners, |
|
||||
dst_partner, context=None): |
|
||||
_logger.debug('_update_foreign_keys for dst_partner: %s for ' |
|
||||
'src_partners: %r', |
|
||||
dst_partner.id, |
|
||||
list(map(operator.attrgetter('id'), src_partners))) |
|
||||
|
|
||||
# find the many2one relation to a partner |
|
||||
proxy = self.pool.get('res.partner') |
|
||||
self.get_fk_on(cr, 'res_partner') |
|
||||
|
|
||||
# ignore two tables |
|
||||
|
|
||||
for table, column in cr.fetchall(): |
|
||||
if 'base_partner_merge_' in table: |
|
||||
continue |
|
||||
partner_ids = tuple(map(int, src_partners)) |
|
||||
|
|
||||
query = ("SELECT column_name FROM information_schema.columns" |
|
||||
" WHERE table_name LIKE '%s'") % (table) |
|
||||
cr.execute(query, ()) |
|
||||
columns = [] |
|
||||
for data in cr.fetchall(): |
|
||||
if data[0] != column: |
|
||||
columns.append(data[0]) |
|
||||
|
|
||||
query_dic = { |
|
||||
'table': table, |
|
||||
'column': column, |
|
||||
'value': columns[0], |
|
||||
} |
|
||||
if len(columns) <= 1: |
|
||||
# unique key treated |
|
||||
query = """ |
|
||||
UPDATE "%(table)s" as ___tu |
|
||||
SET %(column)s = %%s |
|
||||
WHERE |
|
||||
%(column)s = %%s AND |
|
||||
NOT EXISTS ( |
|
||||
SELECT 1 |
|
||||
FROM "%(table)s" as ___tw |
|
||||
WHERE |
|
||||
%(column)s = %%s AND |
|
||||
___tu.%(value)s = ___tw.%(value)s |
|
||||
)""" % query_dic |
|
||||
for partner_id in partner_ids: |
|
||||
cr.execute(query, (dst_partner.id, partner_id, |
|
||||
dst_partner.id)) |
|
||||
else: |
|
||||
cr.execute("SAVEPOINT recursive_partner_savepoint") |
|
||||
try: |
|
||||
query = ('UPDATE "%(table)s" SET %(column)s = %%s WHERE ' |
|
||||
'%(column)s IN %%s') % query_dic |
|
||||
cr.execute(query, (dst_partner.id, partner_ids,)) |
|
||||
|
|
||||
if (column == proxy._parent_name and |
|
||||
table == 'res_partner'): |
|
||||
query = """ |
|
||||
WITH RECURSIVE cycle(id, parent_id) AS ( |
|
||||
SELECT id, parent_id FROM res_partner |
|
||||
UNION |
|
||||
SELECT cycle.id, res_partner.parent_id |
|
||||
FROM res_partner, cycle |
|
||||
WHERE res_partner.id = cycle.parent_id |
|
||||
AND cycle.id != cycle.parent_id |
|
||||
) |
|
||||
SELECT id FROM cycle |
|
||||
WHERE id = parent_id AND id = %s |
|
||||
""" |
|
||||
cr.execute(query, (dst_partner.id,)) |
|
||||
if cr.fetchall(): |
|
||||
cr.execute("ROLLBACK TO SAVEPOINT " |
|
||||
"recursive_partner_savepoint") |
|
||||
finally: |
|
||||
cr.execute("RELEASE SAVEPOINT " |
|
||||
"recursive_partner_savepoint") |
|
||||
|
|
||||
def _update_reference_fields(self, cr, uid, src_partners, dst_partner, |
|
||||
context=None): |
|
||||
_logger.debug('_update_reference_fields for dst_partner: %s for ' |
|
||||
'src_partners: %r', |
|
||||
dst_partner.id, |
|
||||
list(map(operator.attrgetter('id'), src_partners))) |
|
||||
|
|
||||
def update_records(model, src, field_model='model', field_id='res_id', |
|
||||
context=None): |
|
||||
proxy = self.pool.get(model) |
|
||||
if proxy is None: |
|
||||
return |
|
||||
domain = [(field_model, '=', 'res.partner'), |
|
||||
(field_id, '=', src.id)] |
|
||||
ids = proxy.search(cr, openerp.SUPERUSER_ID, |
|
||||
domain, context=context) |
|
||||
if model == 'mail.followers': |
|
||||
# mail.followers have a set semantic |
|
||||
# unlink records that whould trigger a duplicate constraint |
|
||||
# on rewrite |
|
||||
src_objs = proxy.browse(cr, openerp.SUPERUSER_ID, |
|
||||
ids) |
|
||||
target_domain = [(field_model, '=', 'res.partner'), |
|
||||
(field_id, '=', dst_partner.id)] |
|
||||
target_ids = proxy.search(cr, openerp.SUPERUSER_ID, |
|
||||
target_domain, context=context) |
|
||||
dst_followers = proxy.browse(cr, openerp.SUPERUSER_ID, |
|
||||
target_ids).mapped('partner_id') |
|
||||
to_unlink = src_objs.filtered(lambda obj: |
|
||||
obj.partner_id in dst_followers) |
|
||||
to_rewrite = src_objs - to_unlink |
|
||||
to_unlink.unlink() |
|
||||
ids = to_rewrite.ids |
|
||||
|
|
||||
return proxy.write(cr, openerp.SUPERUSER_ID, ids, |
|
||||
{field_id: dst_partner.id}, context=context) |
|
||||
|
|
||||
update_records = functools.partial(update_records, context=context) |
|
||||
|
|
||||
for partner in src_partners: |
|
||||
update_records('base.calendar', src=partner, |
|
||||
field_model='model_id.model') |
|
||||
update_records('ir.attachment', src=partner, |
|
||||
field_model='res_model') |
|
||||
update_records('mail.followers', src=partner, |
|
||||
field_model='res_model') |
|
||||
update_records('mail.message', src=partner) |
|
||||
update_records('marketing.campaign.workitem', src=partner, |
|
||||
field_model='object_id.model') |
|
||||
update_records('ir.model.data', src=partner) |
|
||||
|
|
||||
proxy = self.pool['ir.model.fields'] |
|
||||
domain = [('ttype', '=', 'reference')] |
|
||||
record_ids = proxy.search(cr, openerp.SUPERUSER_ID, domain, |
|
||||
context=context) |
|
||||
|
|
||||
for record in proxy.browse(cr, openerp.SUPERUSER_ID, record_ids, |
|
||||
context=context): |
|
||||
try: |
|
||||
proxy_model = self.pool[record.model] |
|
||||
except KeyError: |
|
||||
# ignore old tables |
|
||||
continue |
|
||||
|
|
||||
if record.model == 'ir.property': |
|
||||
continue |
|
||||
|
|
||||
legacy = proxy_model._columns.get(record.name) |
|
||||
field_spec = proxy_model._fields.get(record.name) |
|
||||
|
|
||||
if not legacy or isinstance(legacy, fields.function) \ |
|
||||
or field_spec.compute: |
|
||||
continue |
|
||||
|
|
||||
for partner in src_partners: |
|
||||
domain = [ |
|
||||
(record.name, '=', 'res.partner,%d' % partner.id) |
|
||||
] |
|
||||
model_ids = proxy_model.search(cr, openerp.SUPERUSER_ID, |
|
||||
domain, context=context) |
|
||||
values = { |
|
||||
record.name: 'res.partner,%d' % dst_partner.id, |
|
||||
} |
|
||||
proxy_model.write(cr, openerp.SUPERUSER_ID, model_ids, values, |
|
||||
context=context) |
|
||||
|
|
||||
def _update_values(self, cr, uid, src_partners, dst_partner, context=None): |
|
||||
_logger.debug('_update_values for dst_partner: %s for src_partners: ' |
|
||||
'%r', |
|
||||
dst_partner.id, |
|
||||
list(map(operator.attrgetter('id'), src_partners))) |
|
||||
|
|
||||
columns = dst_partner._columns |
|
||||
|
|
||||
def write_serializer(column, item): |
|
||||
if isinstance(item, browse_record): |
|
||||
return item.id |
|
||||
else: |
|
||||
return item |
|
||||
|
|
||||
values = dict() |
|
||||
for column, field in columns.iteritems(): |
|
||||
if (field._type not in ('many2many', 'one2many') and |
|
||||
not isinstance(field, fields.function)): |
|
||||
for item in itertools.chain(src_partners, [dst_partner]): |
|
||||
if item[column]: |
|
||||
values[column] = write_serializer(column, |
|
||||
item[column]) |
|
||||
|
|
||||
values.pop('id', None) |
|
||||
parent_id = values.pop('parent_id', None) |
|
||||
dst_partner.write(values) |
|
||||
if parent_id and parent_id != dst_partner.id: |
|
||||
try: |
|
||||
dst_partner.write({'parent_id': parent_id}) |
|
||||
except except_orm: |
|
||||
_logger.info('Skip recursive partner hierarchies for ' |
|
||||
'parent_id %s of partner: %s', |
|
||||
parent_id, dst_partner.id) |
|
||||
|
|
||||
@mute_logger('openerp.osv.expression', 'openerp.osv.orm') |
|
||||
def _merge(self, cr, uid, partner_ids, dst_partner=None, context=None): |
|
||||
proxy = self.pool.get('res.partner') |
|
||||
|
|
||||
partner_ids = proxy.exists(cr, uid, list(partner_ids), |
|
||||
context=context) |
|
||||
if len(partner_ids) < 2: |
|
||||
return |
|
||||
|
|
||||
if len(partner_ids) > 3: |
|
||||
raise except_orm( |
|
||||
_('Error'), |
|
||||
_("For safety reasons, you cannot merge more than 3 contacts " |
|
||||
"together. You can re-open the wizard several times if " |
|
||||
"needed.")) |
|
||||
|
|
||||
if (openerp.SUPERUSER_ID != uid and |
|
||||
len(set(partner.email for partner |
|
||||
in proxy.browse(cr, uid, partner_ids, |
|
||||
context=context))) > 1): |
|
||||
raise except_orm( |
|
||||
_('Error'), |
|
||||
_("All contacts must have the same email. Only the " |
|
||||
"Administrator can merge contacts with different emails.")) |
|
||||
|
|
||||
if dst_partner and dst_partner.id in partner_ids: |
|
||||
src_partners = proxy.browse(cr, uid, |
|
||||
[id for id in partner_ids |
|
||||
if id != dst_partner.id], |
|
||||
context=context) |
|
||||
else: |
|
||||
ordered_partners = self._get_ordered_partner(cr, uid, partner_ids, |
|
||||
context) |
|
||||
dst_partner = ordered_partners[-1] |
|
||||
src_partners = ordered_partners[:-1] |
|
||||
_logger.info("dst_partner: %s", dst_partner.id) |
|
||||
|
|
||||
if (openerp.SUPERUSER_ID != uid and |
|
||||
self._model_is_installed( |
|
||||
cr, uid, 'account.move.line', context=context) and |
|
||||
self.pool['account.move.line'].search( |
|
||||
cr, openerp.SUPERUSER_ID, |
|
||||
[('partner_id', 'in', [partner.id for partner |
|
||||
in src_partners])], |
|
||||
context=context)): |
|
||||
raise except_orm( |
|
||||
_('Error'), |
|
||||
_("Only the destination contact may be linked to existing " |
|
||||
"Journal Items. Please ask the Administrator if you need to" |
|
||||
" merge several contacts linked to existing Journal " |
|
||||
"Items.")) |
|
||||
self._update_foreign_keys( |
|
||||
cr, uid, src_partners, dst_partner, context=context) |
|
||||
self._update_reference_fields( |
|
||||
cr, uid, src_partners, dst_partner, context=context) |
|
||||
self._update_values( |
|
||||
cr, uid, src_partners, dst_partner, context=context) |
|
||||
_logger.info('(uid = %s) merged the partners %r with %s', |
|
||||
uid, |
|
||||
list(map(operator.attrgetter('id'), src_partners)), |
|
||||
dst_partner.id) |
|
||||
dst_partner.message_post( |
|
||||
body='%s %s' % ( |
|
||||
_("Merged with the following partners:"), |
|
||||
", ".join( |
|
||||
'%s<%s>(ID %s)' % (p.name, p.email or 'n/a', p.id) |
|
||||
for p in src_partners |
|
||||
) |
|
||||
) |
|
||||
) |
|
||||
|
|
||||
for partner in src_partners: |
|
||||
partner.unlink() |
|
||||
|
|
||||
def clean_emails(self, cr, uid, context=None): |
|
||||
""" |
|
||||
Clean the email address of the partner, if there is an email field |
|
||||
with a minimum of two addresses, the system will create a new partner, |
|
||||
with the information of the previous one and will copy the new cleaned |
|
||||
email into the email field. |
|
||||
""" |
|
||||
if context is None: |
|
||||
context = {} |
|
||||
|
|
||||
proxy_model = self.pool['ir.model.fields'] |
|
||||
field_ids = proxy_model.search(cr, uid, |
|
||||
[('model', '=', 'res.partner'), |
|
||||
('ttype', 'like', '%2many')], |
|
||||
context=context) |
|
||||
fields = proxy_model.read(cr, uid, field_ids, context=context) |
|
||||
reset_fields = dict((field['name'], []) for field in fields) |
|
||||
|
|
||||
proxy_partner = self.pool['res.partner'] |
|
||||
context['active_test'] = False |
|
||||
ids = proxy_partner.search(cr, uid, [], context=context) |
|
||||
|
|
||||
fields = ['name', 'var' 'partner_id' 'is_company', 'email'] |
|
||||
partners = proxy_partner.read(cr, uid, ids, fields, context=context) |
|
||||
|
|
||||
partners.sort(key=operator.itemgetter('id')) |
|
||||
partners_len = len(partners) |
|
||||
|
|
||||
_logger.info('partner_len: %r', partners_len) |
|
||||
|
|
||||
for idx, partner in enumerate(partners): |
|
||||
if not partner['email']: |
|
||||
continue |
|
||||
|
|
||||
percent = (idx / float(partners_len)) * 100.0 |
|
||||
_logger.info('idx: %r', idx) |
|
||||
_logger.info('percent: %r', percent) |
|
||||
try: |
|
||||
emails = sanitize_email(partner['email']) |
|
||||
head, tail = emails[:1], emails[1:] |
|
||||
email = head[0] if head else False |
|
||||
|
|
||||
proxy_partner.write(cr, uid, [partner['id']], |
|
||||
{'email': email}, context=context) |
|
||||
|
|
||||
for email in tail: |
|
||||
values = dict(reset_fields, email=email) |
|
||||
proxy_partner.copy(cr, uid, partner['id'], values, |
|
||||
context=context) |
|
||||
|
|
||||
except Exception: |
|
||||
_logger.exception("There is a problem with this partner: %r", |
|
||||
partner) |
|
||||
raise |
|
||||
return True |
|
||||
|
|
||||
def close_cb(self, cr, uid, ids, context=None): |
|
||||
return {'type': 'ir.actions.act_window_close'} |
|
||||
|
|
||||
def _generate_query(self, fields, maximum_group=100): |
|
||||
group_fields = ', '.join(fields) |
|
||||
|
|
||||
filters = [] |
|
||||
for field in fields: |
|
||||
if field in ['email', 'name']: |
|
||||
filters.append((field, 'IS NOT', 'NULL')) |
|
||||
|
|
||||
criteria = ' AND '.join('%s %s %s' % (field, operator, value) |
|
||||
for field, operator, value in filters) |
|
||||
|
|
||||
text = [ |
|
||||
"SELECT min(id), array_agg(id)", |
|
||||
"FROM res_partner", |
|
||||
] |
|
||||
|
|
||||
if criteria: |
|
||||
text.append('WHERE %s' % criteria) |
|
||||
|
|
||||
text.extend([ |
|
||||
"GROUP BY %s" % group_fields, |
|
||||
"HAVING COUNT(*) >= 2", |
|
||||
"ORDER BY min(id)", |
|
||||
]) |
|
||||
|
|
||||
if maximum_group: |
|
||||
text.extend([ |
|
||||
"LIMIT %s" % maximum_group, |
|
||||
]) |
|
||||
|
|
||||
return ' '.join(text) |
|
||||
|
|
||||
def _compute_selected_groupby(self, this): |
|
||||
group_by_str = 'group_by_' |
|
||||
group_by_len = len(group_by_str) |
|
||||
|
|
||||
fields = [ |
|
||||
key[group_by_len:] |
|
||||
for key in self._columns.keys() |
|
||||
if key.startswith(group_by_str) |
|
||||
] |
|
||||
|
|
||||
groups = [ |
|
||||
field |
|
||||
for field in fields |
|
||||
if getattr(this, '%s%s' % (group_by_str, field), False) |
|
||||
] |
|
||||
|
|
||||
if not groups: |
|
||||
raise except_orm(_('Error'), |
|
||||
_("You have to specify a filter for your " |
|
||||
"selection")) |
|
||||
|
|
||||
return groups |
|
||||
|
|
||||
def next_cb(self, cr, uid, ids, context=None): |
|
||||
""" |
|
||||
Don't compute any thing |
|
||||
""" |
|
||||
context = dict(context or {}, active_test=False) |
|
||||
this = self.browse(cr, uid, ids[0], context=context) |
|
||||
if this.current_line_id: |
|
||||
this.current_line_id.unlink() |
|
||||
return self._next_screen(cr, uid, this, context) |
|
||||
|
|
||||
def _get_ordered_partner(self, cr, uid, partner_ids, context=None): |
|
||||
partners = self.pool.get('res.partner' |
|
||||
).browse(cr, uid, |
|
||||
list(partner_ids), |
|
||||
context=context) |
|
||||
ordered_partners = sorted( |
|
||||
sorted( |
|
||||
partners, |
|
||||
key=operator.attrgetter('create_date'), |
|
||||
reverse=True |
|
||||
), |
|
||||
key=operator.attrgetter('active'), |
|
||||
reverse=True |
|
||||
) |
|
||||
return ordered_partners |
|
||||
|
|
||||
def _next_screen(self, cr, uid, this, context=None): |
|
||||
this.refresh() |
|
||||
values = {} |
|
||||
if this.line_ids: |
|
||||
# in this case, we try to find the next record. |
|
||||
current_line = this.line_ids[0] |
|
||||
current_partner_ids = literal_eval(current_line.aggr_ids) |
|
||||
values.update({ |
|
||||
'current_line_id': current_line.id, |
|
||||
'partner_ids': [(6, 0, current_partner_ids)], |
|
||||
'dst_partner_id': self._get_ordered_partner( |
|
||||
cr, uid, |
|
||||
current_partner_ids, |
|
||||
context |
|
||||
)[-1].id, |
|
||||
'state': 'selection', |
|
||||
}) |
|
||||
else: |
|
||||
values.update({ |
|
||||
'current_line_id': False, |
|
||||
'partner_ids': [], |
|
||||
'state': 'finished', |
|
||||
}) |
|
||||
|
|
||||
this.write(values) |
|
||||
|
|
||||
return { |
|
||||
'type': 'ir.actions.act_window', |
|
||||
'res_model': this._name, |
|
||||
'res_id': this.id, |
|
||||
'view_mode': 'form', |
|
||||
'target': 'new', |
|
||||
} |
|
||||
|
|
||||
def _model_is_installed(self, cr, uid, model, context=None): |
|
||||
proxy = self.pool.get('ir.model') |
|
||||
domain = [('model', '=', model)] |
|
||||
return proxy.search_count(cr, uid, domain, context=context) > 0 |
|
||||
|
|
||||
def _partner_use_in(self, cr, uid, aggr_ids, models, context=None): |
|
||||
""" |
|
||||
Check if there is no occurence of this group of partner in the selected |
|
||||
model |
|
||||
""" |
|
||||
for model, field in models.iteritems(): |
|
||||
proxy = self.pool.get(model) |
|
||||
domain = [(field, 'in', aggr_ids)] |
|
||||
if proxy.search_count(cr, uid, domain, context=context): |
|
||||
return True |
|
||||
return False |
|
||||
|
|
||||
def compute_models(self, cr, uid, ids, context=None): |
|
||||
""" |
|
||||
Compute the different models needed by the system if you want to |
|
||||
exclude some partners. |
|
||||
""" |
|
||||
assert is_integer_list(ids) |
|
||||
|
|
||||
this = self.browse(cr, uid, ids[0], context=context) |
|
||||
|
|
||||
models = {} |
|
||||
if this.exclude_contact: |
|
||||
models['res.users'] = 'partner_id' |
|
||||
|
|
||||
if (self._model_is_installed( |
|
||||
cr, uid, 'account.move.line', context=context) and |
|
||||
this.exclude_journal_item): |
|
||||
models['account.move.line'] = 'partner_id' |
|
||||
|
|
||||
return models |
|
||||
|
|
||||
def _process_query(self, cr, uid, ids, query, context=None): |
|
||||
""" |
|
||||
Execute the select request and write the result in this wizard |
|
||||
""" |
|
||||
proxy = self.pool.get('base.partner.merge.line') |
|
||||
this = self.browse(cr, uid, ids[0], context=context) |
|
||||
models = self.compute_models(cr, uid, ids, context=context) |
|
||||
cr.execute(query) |
|
||||
|
|
||||
counter = 0 |
|
||||
for min_id, aggr_ids in cr.fetchall(): |
|
||||
if models and self._partner_use_in(cr, uid, aggr_ids, models, |
|
||||
context=context): |
|
||||
continue |
|
||||
values = { |
|
||||
'wizard_id': this.id, |
|
||||
'min_id': min_id, |
|
||||
'aggr_ids': aggr_ids, |
|
||||
} |
|
||||
|
|
||||
proxy.create(cr, uid, values, context=context) |
|
||||
counter += 1 |
|
||||
|
|
||||
values = { |
|
||||
'state': 'selection', |
|
||||
'number_group': counter, |
|
||||
} |
|
||||
|
|
||||
this.write(values) |
|
||||
|
|
||||
_logger.info("counter: %s", counter) |
|
||||
|
|
||||
def start_process_cb(self, cr, uid, ids, context=None): |
|
||||
""" |
|
||||
Start the process. |
|
||||
* Compute the selected groups (with duplication) |
|
||||
* If the user has selected the 'exclude_XXX' fields, avoid the |
|
||||
partners. |
|
||||
""" |
|
||||
assert is_integer_list(ids) |
|
||||
|
|
||||
context = dict(context or {}, active_test=False) |
|
||||
this = self.browse(cr, uid, ids[0], context=context) |
|
||||
groups = self._compute_selected_groupby(this) |
|
||||
query = self._generate_query(groups, this.maximum_group) |
|
||||
self._process_query(cr, uid, ids, query, context=context) |
|
||||
|
|
||||
return self._next_screen(cr, uid, this, context) |
|
||||
|
|
||||
def automatic_process_cb(self, cr, uid, ids, context=None): |
|
||||
assert is_integer_list(ids) |
|
||||
this = self.browse(cr, uid, ids[0], context=context) |
|
||||
this.start_process_cb() |
|
||||
this.refresh() |
|
||||
|
|
||||
for line in this.line_ids: |
|
||||
partner_ids = literal_eval(line.aggr_ids) |
|
||||
self._merge(cr, uid, partner_ids, context=context) |
|
||||
line.unlink() |
|
||||
cr.commit() |
|
||||
|
|
||||
this.write({'state': 'finished'}) |
|
||||
return { |
|
||||
'type': 'ir.actions.act_window', |
|
||||
'res_model': this._name, |
|
||||
'res_id': this.id, |
|
||||
'view_mode': 'form', |
|
||||
'target': 'new', |
|
||||
} |
|
||||
|
|
||||
def parent_migration_process_cb(self, cr, uid, ids, context=None): |
|
||||
assert is_integer_list(ids) |
|
||||
|
|
||||
context = dict(context or {}, active_test=False) |
|
||||
this = self.browse(cr, uid, ids[0], context=context) |
|
||||
|
|
||||
query = """ |
|
||||
SELECT |
|
||||
min(p1.id), |
|
||||
array_agg(DISTINCT p1.id) |
|
||||
FROM |
|
||||
res_partner as p1 |
|
||||
INNER join |
|
||||
res_partner as p2 |
|
||||
ON |
|
||||
p1.email = p2.email AND |
|
||||
p1.name = p2.name AND |
|
||||
(p1.parent_id = p2.id OR p1.id = p2.parent_id) |
|
||||
WHERE |
|
||||
p2.id IS NOT NULL |
|
||||
GROUP BY |
|
||||
p1.email, |
|
||||
p1.name, |
|
||||
CASE WHEN p1.parent_id = p2.id THEN p2.id |
|
||||
ELSE p1.id |
|
||||
END |
|
||||
HAVING COUNT(*) >= 2 |
|
||||
ORDER BY |
|
||||
min(p1.id) |
|
||||
""" |
|
||||
|
|
||||
self._process_query(cr, uid, ids, query, context=context) |
|
||||
|
|
||||
for line in this.line_ids: |
|
||||
partner_ids = literal_eval(line.aggr_ids) |
|
||||
self._merge(cr, uid, partner_ids, context=context) |
|
||||
line.unlink() |
|
||||
cr.commit() |
|
||||
|
|
||||
this.write({'state': 'finished'}) |
|
||||
|
|
||||
cr.execute(""" |
|
||||
UPDATE |
|
||||
res_partner |
|
||||
SET |
|
||||
is_company = NULL, |
|
||||
parent_id = NULL |
|
||||
WHERE |
|
||||
parent_id = id |
|
||||
""") |
|
||||
|
|
||||
return { |
|
||||
'type': 'ir.actions.act_window', |
|
||||
'res_model': this._name, |
|
||||
'res_id': this.id, |
|
||||
'view_mode': 'form', |
|
||||
'target': 'new', |
|
||||
} |
|
||||
|
|
||||
def update_all_process_cb(self, cr, uid, ids, context=None): |
|
||||
assert is_integer_list(ids) |
|
||||
|
|
||||
# WITH RECURSIVE cycle(id, parent_id) AS ( |
|
||||
# SELECT id, parent_id FROM res_partner |
|
||||
# UNION |
|
||||
# SELECT cycle.id, res_partner.parent_id |
|
||||
# FROM res_partner, cycle |
|
||||
# WHERE res_partner.id = cycle.parent_id AND |
|
||||
# cycle.id != cycle.parent_id |
|
||||
# ) |
|
||||
# UPDATE res_partner |
|
||||
# SET parent_id = NULL |
|
||||
# WHERE id in (SELECT id FROM cycle WHERE id = parent_id); |
|
||||
|
|
||||
this = self.browse(cr, uid, ids[0], context=context) |
|
||||
|
|
||||
self.parent_migration_process_cb(cr, uid, ids, context=None) |
|
||||
|
|
||||
list_merge = [ |
|
||||
{'group_by_vat': True, |
|
||||
'group_by_email': True, |
|
||||
'group_by_name': True}, |
|
||||
# {'group_by_name': True, |
|
||||
# 'group_by_is_company': True, |
|
||||
# 'group_by_parent_id': True}, |
|
||||
# {'group_by_email': True, |
|
||||
# 'group_by_is_company': True, |
|
||||
# 'group_by_parent_id': True}, |
|
||||
# {'group_by_name': True, |
|
||||
# 'group_by_vat': True, |
|
||||
# 'group_by_is_company': True, |
|
||||
# 'exclude_journal_item': True}, |
|
||||
# {'group_by_email': True, |
|
||||
# 'group_by_vat': True, |
|
||||
# 'group_by_is_company': True, |
|
||||
# 'exclude_journal_item': True}, |
|
||||
# {'group_by_email': True, |
|
||||
# 'group_by_is_company': True, |
|
||||
# 'exclude_contact': True, |
|
||||
# 'exclude_journal_item': True}, |
|
||||
# {'group_by_name': True, |
|
||||
# 'group_by_is_company': True, |
|
||||
# 'exclude_contact': True, |
|
||||
# 'exclude_journal_item': True} |
|
||||
] |
|
||||
|
|
||||
for merge_value in list_merge: |
|
||||
id = self.create(cr, uid, merge_value, context=context) |
|
||||
self.automatic_process_cb(cr, uid, [id], context=context) |
|
||||
|
|
||||
cr.execute(""" |
|
||||
UPDATE |
|
||||
res_partner |
|
||||
SET |
|
||||
is_company = NULL |
|
||||
WHERE |
|
||||
parent_id IS NOT NULL AND |
|
||||
is_company IS NOT NULL |
|
||||
""") |
|
||||
|
|
||||
# cr.execute(""" |
|
||||
# UPDATE |
|
||||
# res_partner as p1 |
|
||||
# SET |
|
||||
# is_company = NULL, |
|
||||
# parent_id = ( |
|
||||
# SELECT p2.id |
|
||||
# FROM res_partner as p2 |
|
||||
# WHERE p2.email = p1.email AND |
|
||||
# p2.parent_id != p2.id |
|
||||
# LIMIT 1 |
|
||||
# ) |
|
||||
# WHERE |
|
||||
# p1.parent_id = p1.id |
|
||||
# """) |
|
||||
|
|
||||
return self._next_screen(cr, uid, this, context) |
|
||||
|
|
||||
def merge_cb(self, cr, uid, ids, context=None): |
|
||||
assert is_integer_list(ids) |
|
||||
|
|
||||
context = dict(context or {}, active_test=False) |
|
||||
this = self.browse(cr, uid, ids[0], context=context) |
|
||||
|
|
||||
partner_ids = set(map(int, this.partner_ids)) |
|
||||
if not partner_ids: |
|
||||
this.write({'state': 'finished'}) |
|
||||
return { |
|
||||
'type': 'ir.actions.act_window', |
|
||||
'res_model': this._name, |
|
||||
'res_id': this.id, |
|
||||
'view_mode': 'form', |
|
||||
'target': 'new', |
|
||||
} |
|
||||
|
|
||||
self._merge(cr, uid, partner_ids, this.dst_partner_id, |
|
||||
context=context) |
|
||||
|
|
||||
if this.current_line_id: |
|
||||
this.current_line_id.unlink() |
|
||||
|
|
||||
return self._next_screen(cr, uid, this, context) |
|
||||
|
|
||||
def auto_set_parent_id(self, cr, uid, ids, context=None): |
|
||||
assert is_integer_list(ids) |
|
||||
|
|
||||
# select partner who have one least invoice |
|
||||
partner_treated = ['@gmail.com'] |
|
||||
cr.execute(""" SELECT p.id, p.email |
|
||||
FROM res_partner as p |
|
||||
LEFT JOIN account_invoice as a |
|
||||
ON p.id = a.partner_id AND a.state in ('open','paid') |
|
||||
WHERE p.grade_id is NOT NULL |
|
||||
GROUP BY p.id |
|
||||
ORDER BY COUNT(a.id) DESC |
|
||||
""") |
|
||||
re_email = re.compile(r".*@") |
|
||||
for id, email in cr.fetchall(): |
|
||||
# check email domain |
|
||||
email = re_email.sub("@", email or "") |
|
||||
if not email or email in partner_treated: |
|
||||
continue |
|
||||
partner_treated.append(email) |
|
||||
|
|
||||
# don't update the partners if they are more of one who have |
|
||||
# invoice |
|
||||
cr.execute(""" |
|
||||
SELECT * |
|
||||
FROM res_partner as p |
|
||||
WHERE p.id != %s AND p.email LIKE '%%%s' AND |
|
||||
EXISTS (SELECT * FROM account_invoice as a |
|
||||
WHERE p.id = a.partner_id |
|
||||
AND a.state in ('open','paid')) |
|
||||
""" % (id, email)) |
|
||||
|
|
||||
if len(cr.fetchall()) > 1: |
|
||||
_logger.info("%s MORE OF ONE COMPANY", email) |
|
||||
continue |
|
||||
|
|
||||
# to display changed values |
|
||||
cr.execute(""" SELECT id,email |
|
||||
FROM res_partner |
|
||||
WHERE parent_id != %s |
|
||||
AND id != %s AND email LIKE '%%%s' |
|
||||
""" % (id, id, email)) |
|
||||
_logger.info("%r", cr.fetchall()) |
|
||||
|
|
||||
# upgrade |
|
||||
cr.execute(""" UPDATE res_partner |
|
||||
SET parent_id = %s |
|
||||
WHERE id != %s AND email LIKE '%%%s' |
|
||||
""" % (id, id, email)) |
|
||||
return False |
|
@ -0,0 +1,29 @@ |
|||||
|
# -*- coding: utf-8 -*- |
||||
|
# © 2017 Therp BV <http://therp.nl> |
||||
|
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html). |
||||
|
import inspect |
||||
|
from openerp.models import MetaModel |
||||
|
|
||||
|
|
||||
|
def post_load_hook(): |
||||
|
""" |
||||
|
We try to be smart here: If the crm module is to be loaded too |
||||
|
(or is already loaded), we remove our own models again in order not to |
||||
|
clash with the CRM ones: https://github.com/OCA/partner-contact/issues/283 |
||||
|
""" |
||||
|
for frame, filename, lineno, funcname, line, index in inspect.stack(): |
||||
|
# walk up the stack until we're in load_module_graph |
||||
|
if 'graph' in frame.f_locals: |
||||
|
graph = frame.f_locals['graph'] |
||||
|
package = frame.f_locals['package'] |
||||
|
if any(p.name == 'crm' for p in graph): |
||||
|
# so crm is installed, then we need to remove the models |
||||
|
# we pull from CRM again |
||||
|
MetaModel.module_to_models['base_partner_merge'] = [ |
||||
|
cls |
||||
|
for cls in MetaModel.module_to_models['base_partner_merge'] |
||||
|
if 'NoCRM' not in cls.__name__ |
||||
|
] |
||||
|
# and in this case, we also don't want to load our xml file |
||||
|
package.data['data'].remove('views/base_partner_merge.xml') |
||||
|
break |
@ -0,0 +1,5 @@ |
|||||
|
# -*- coding: utf-8 -*- |
||||
|
|
||||
|
from . import base_partner_merge # noqa |
||||
|
from . import res_partner |
||||
|
from . import base_partner_merge_automatic_wizard |
@ -0,0 +1,16 @@ |
|||||
|
# -*- coding: utf-8 -*- |
||||
|
# © 2016 Camptocamp SA |
||||
|
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). |
||||
|
from openerp.addons.crm.base_partner_merge import * # noqa |
||||
|
|
||||
|
|
||||
|
class NoCRMResPartner(ResPartner): # noqa |
||||
|
_module = 'base_partner_merge' |
||||
|
|
||||
|
|
||||
|
class NoCRMMergePartnerLine(MergePartnerLine): # noqa |
||||
|
_module = 'base_partner_merge' |
||||
|
|
||||
|
|
||||
|
class NoCRMMergePartnerAutomatic(MergePartnerAutomatic): # noqa |
||||
|
_module = 'base_partner_merge' |
@ -0,0 +1,30 @@ |
|||||
|
# -*- coding: utf-8 -*- |
||||
|
# © 2017 Sunflower IT <http://sunflowerweb.nl> |
||||
|
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html). |
||||
|
|
||||
|
from openerp import fields, models, api |
||||
|
from ast import literal_eval |
||||
|
|
||||
|
|
||||
|
class MergePartnerAutomatic(models.TransientModel): |
||||
|
_inherit = 'base.partner.merge.automatic.wizard' |
||||
|
|
||||
|
# Enable deduplicating by reference |
||||
|
group_by_ref = fields.Boolean('Reference') |
||||
|
|
||||
|
@api.multi |
||||
|
def _process_query(self, query): |
||||
|
ret = super(MergePartnerAutomatic, self)._process_query(query) |
||||
|
|
||||
|
# If 'extra_domain', deduplicate only the records matching the domain |
||||
|
extra_domain = self.env.context.get('partner_merge_domain', []) |
||||
|
if extra_domain: |
||||
|
for line in self.line_ids: |
||||
|
domain = [('id', 'in', literal_eval(line.aggr_ids))] |
||||
|
domain.extend(extra_domain) |
||||
|
aggr_ids = self.env['res.partner'].search(domain).ids |
||||
|
if len(aggr_ids) > 1: |
||||
|
line.aggr_ids = str(aggr_ids) |
||||
|
else: |
||||
|
line.unlink() |
||||
|
return ret |
@ -0,0 +1,21 @@ |
|||||
|
# -*- coding: utf-8 -*- |
||||
|
# © 2017 Sunflower IT <http://sunflowerweb.nl> |
||||
|
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html). |
||||
|
|
||||
|
from openerp import models, api |
||||
|
|
||||
|
|
||||
|
class ResPartnerChanges(models.Model): |
||||
|
_inherit = 'res.partner' |
||||
|
|
||||
|
@api.model |
||||
|
def deduplicate_on_fields(self, fields_list, domain=None): |
||||
|
""" Merge contacts """ |
||||
|
wizard_obj = self.env['base.partner.merge.automatic.wizard'] |
||||
|
if domain: |
||||
|
wizard_obj = wizard_obj.with_context(partner_merge_domain=domain) |
||||
|
params = {'state': 'option'} |
||||
|
for field in fields_list: |
||||
|
params['group_by_%s' % (field,)] = True |
||||
|
wizard = wizard_obj.create(params) |
||||
|
wizard.automatic_process_cb() |
@ -0,0 +1,3 @@ |
|||||
|
# -*- coding: utf-8 -*- |
||||
|
|
||||
|
from . import test_merge |
@ -0,0 +1,79 @@ |
|||||
|
# -*- coding: utf-8 -*- |
||||
|
# © 2017 Sunflower IT <http://sunflowerweb.nl> |
||||
|
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html). |
||||
|
|
||||
|
from openerp.tests.common import TransactionCase |
||||
|
|
||||
|
|
||||
|
class PartnerMergeTestCase(TransactionCase): |
||||
|
"""Tests for Partner Merge""" |
||||
|
|
||||
|
def setUp(self): |
||||
|
super(PartnerMergeTestCase, self).setUp() |
||||
|
self.partner = self.env['res.partner'] |
||||
|
self.merge_wizard = \ |
||||
|
self.env['base.partner.merge.automatic.wizard'] |
||||
|
self.donald_domain = [('name', '=', 'Donald Duck')] |
||||
|
self.mickey_domain = [('name', '=', 'Mickey Mouse')] |
||||
|
|
||||
|
def _unlink_all(self): |
||||
|
self.partner.search(self.donald_domain).unlink() |
||||
|
self.partner.search(self.mickey_domain).unlink() |
||||
|
|
||||
|
def _count_donalds_mickeys(self, donalds, mickeys): |
||||
|
self.assertEquals( |
||||
|
len(self.partner.search(self.donald_domain)), donalds) |
||||
|
self.assertEquals( |
||||
|
len(self.partner.search(self.mickey_domain)), mickeys) |
||||
|
|
||||
|
def _create_duplicates(self, field1, value1, field2, values2): |
||||
|
for value2 in values2: |
||||
|
self.partner.create({ |
||||
|
field1: value1, |
||||
|
field2: value2, |
||||
|
}) |
||||
|
|
||||
|
def test_10_name_merge(self): |
||||
|
""" Merge users with duplicate names """ |
||||
|
self._unlink_all() |
||||
|
self._create_duplicates( |
||||
|
'name', 'Donald Duck', |
||||
|
'email', ['donald@therp.nl', 'donald@sunflowerweb.nl'], |
||||
|
) |
||||
|
self._create_duplicates( |
||||
|
'name', 'Mickey Mouse', |
||||
|
'email', ['mickey@therp.nl', 'mickey@sunflowerweb.nl'], |
||||
|
) |
||||
|
self._count_donalds_mickeys(2, 2) |
||||
|
# Merge all names that start with 'D', |
||||
|
self.partner.deduplicate_on_fields( |
||||
|
['name'], domain=[('name', 'like', 'D%')], |
||||
|
) |
||||
|
self._count_donalds_mickeys(1, 2) |
||||
|
|
||||
|
def test_20_ref_merge(self): |
||||
|
""" Merge users with duplicate references """ |
||||
|
self._unlink_all() |
||||
|
self._create_duplicates( |
||||
|
'ref', 'DD123', |
||||
|
'name', ['Donald Duck', 'Mickey Mouse'], |
||||
|
) |
||||
|
# Merge on reference, leaving out guys that have no ref |
||||
|
self.partner.deduplicate_on_fields( |
||||
|
['ref'], domain=[('ref', '=', 'DD123')], |
||||
|
) |
||||
|
# Test if only one remains after |
||||
|
partners = self.partner.search([('ref', '=', 'DD123')]) |
||||
|
self.assertEquals(len(partners), 1) |
||||
|
|
||||
|
def test_30_ref_merge(self): |
||||
|
""" Fringe case: three guys, two to merge """ |
||||
|
self._unlink_all() |
||||
|
self._create_duplicates( |
||||
|
'ref', 'DD123', |
||||
|
'name', ['Donald Duck', 'Donald Duck', 'Mickey Mouse'], |
||||
|
) |
||||
|
self.partner.deduplicate_on_fields( |
||||
|
['ref'], domain=[('name', '=', 'Donald Duck')], |
||||
|
) |
||||
|
self._count_donalds_mickeys(1, 1) |
@ -1,135 +1,4 @@ |
|||||
# -*- coding: utf-8 -*- |
# -*- coding: utf-8 -*- |
||||
# RFC 2822 - style email validation for Python |
|
||||
# (c) 2012 Syrus Akbary <me@syrusakbary.com> |
|
||||
# Extended from (c) 2011 Noel Bush <noel@aitools.org> |
|
||||
# for support of mx and user check |
|
||||
# This code is made available to you under the GNU LGPL v3. |
|
||||
# |
|
||||
# This module provides a single method, valid_email_address(), |
|
||||
# which returns True or False to indicate whether a given address |
|
||||
# is valid according to the 'addr-spec' part of the specification |
|
||||
# given in RFC 2822. Ideally, we would like to find this |
|
||||
# in some other library, already thoroughly tested and well- |
|
||||
# maintained. The standard Python library email.utils |
|
||||
# contains a parse_addr() function, but it is not sufficient |
|
||||
# to detect many malformed addresses. |
|
||||
# |
|
||||
# This implementation aims to be faithful to the RFC, with the |
|
||||
# exception of a circular definition (see comments below), and |
|
||||
# with the omission of the pattern components marked as "obsolete". |
|
||||
|
|
||||
import re |
|
||||
import smtplib |
|
||||
|
|
||||
try: |
|
||||
import DNS |
|
||||
ServerError = DNS.ServerError |
|
||||
except: |
|
||||
DNS = None |
|
||||
|
|
||||
class ServerError(Exception): |
|
||||
pass |
|
||||
# All we are really doing is comparing the input string to one |
|
||||
# gigantic regular expression. But building that regexp, and |
|
||||
# ensuring its correctness, is made much easier by assembling it |
|
||||
# from the "tokens" defined by the RFC. Each of these tokens is |
|
||||
# tested in the accompanying unit test file. |
|
||||
# |
|
||||
# The section of RFC 2822 from which each pattern component is |
|
||||
# derived is given in an accompanying comment. |
|
||||
# |
|
||||
# (To make things simple, every string below is given as 'raw', |
|
||||
# even when it's not strictly necessary. This way we don't forget |
|
||||
# when it is necessary.) |
|
||||
# |
|
||||
WSP = r'[ \t]' # see 2.2.2. Structured Header Field Bodies |
|
||||
CRLF = r'(?:\r\n)' # see 2.2.3. Long Header Fields |
|
||||
NO_WS_CTL = r'\x01-\x08\x0b\x0c\x0f-\x1f\x7f' # see 3.2.1. Primitive Tokens |
|
||||
QUOTED_PAIR = r'(?:\\.)' # see 3.2.2. Quoted characters |
|
||||
FWS = r'(?:(?:{0}*{1})?{0}+)'.format(WSP, CRLF) |
|
||||
# see 3.2.3. Folding white space and comments |
|
||||
CTEXT = r'[{0}\x21-\x27\x2a-\x5b\x5d-\x7e]'.format( |
|
||||
NO_WS_CTL) # see 3.2.3 |
|
||||
# see 3.2.3 (NB: The RFC includes COMMENT here as well, but that would be |
|
||||
# circular.) |
|
||||
CCONTENT = r'(?:{0}|{1})'.format(CTEXT, QUOTED_PAIR) |
|
||||
COMMENT = r'\((?:{0}?{1})*{0}?\)'.format( |
|
||||
FWS, CCONTENT) # see 3.2.3 |
|
||||
CFWS = r'(?:{0}?{1})*(?:{0}?{1}|{0})'.format( |
|
||||
FWS, COMMENT) # see 3.2.3 |
|
||||
ATEXT = r'[\w!#$%&\'\*\+\-/=\?\^`\{\|\}~]' # see 3.2.4. Atom |
|
||||
ATOM = r'{0}?{1}+{0}?'.format(CFWS, ATEXT) |
|
||||
# see 3.2.4 |
|
||||
DOT_ATOM_TEXT = r'{0}+(?:\.{0}+)*'.format( |
|
||||
ATEXT) # see 3.2.4 |
|
||||
DOT_ATOM = r'{0}?{1}{0}?'.format( |
|
||||
CFWS, DOT_ATOM_TEXT) # see 3.2.4 |
|
||||
QTEXT = r'[{0}\x21\x23-\x5b\x5d-\x7e]'.format( |
|
||||
NO_WS_CTL) # see 3.2.5. Quoted strings |
|
||||
QCONTENT = r'(?:{0}|{1})'.format(QTEXT, QUOTED_PAIR) |
|
||||
# see 3.2.5 |
|
||||
QUOTED_STRING = r'{0}?"(?:{1}?{2})*{1}?"{0}?'.format(CFWS, FWS, QCONTENT) |
|
||||
LOCAL_PART = r'(?:{0}|{1})'.format(DOT_ATOM, QUOTED_STRING) |
|
||||
# see 3.4.1. Addr-spec specification |
|
||||
DTEXT = r'[{0}\x21-\x5a\x5e-\x7e]'.format( |
|
||||
NO_WS_CTL) # see 3.4.1 |
|
||||
DCONTENT = r'(?:{0}|{1})'.format(DTEXT, QUOTED_PAIR) |
|
||||
# see 3.4.1 |
|
||||
DOMAIN_LITERAL = r'{0}?\[(?:{1}?{2})*{1}?\]{0}?'.format( |
|
||||
CFWS, FWS, DCONTENT) # see 3.4.1 |
|
||||
DOMAIN = r'(?:{0}|{1})'.format(DOT_ATOM, DOMAIN_LITERAL) |
|
||||
# see 3.4.1 |
|
||||
ADDR_SPEC = r'{0}@{1}'.format( |
|
||||
LOCAL_PART, DOMAIN) # see 3.4.1 |
|
||||
|
|
||||
# A valid address will match exactly the 3.4.1 addr-spec. |
|
||||
VALID_ADDRESS_REGEXP = '^' + ADDR_SPEC + '$' |
|
||||
|
|
||||
|
|
||||
def validate_email(email, check_mx=False, verify=False): |
|
||||
"""Indicate whether the given string is a valid email address |
|
||||
according to the 'addr-spec' portion of RFC 2822 (see section |
|
||||
3.4.1). Parts of the spec that are marked obsolete are *not* |
|
||||
included in this test, and certain arcane constructions that |
|
||||
depend on circular definitions in the spec may not pass, but in |
|
||||
general this should correctly identify any email address likely |
|
||||
to be in use as of 2011.""" |
|
||||
try: |
|
||||
assert re.match(VALID_ADDRESS_REGEXP, email) is not None |
|
||||
check_mx |= verify |
|
||||
if check_mx: |
|
||||
if not DNS: |
|
||||
raise Exception('For check the mx records or check if the ' |
|
||||
'email exists you must have installed pyDNS ' |
|
||||
'python package') |
|
||||
DNS.DiscoverNameServers() |
|
||||
hostname = email[email.find('@') + 1:] |
|
||||
mx_hosts = DNS.mxlookup(hostname) |
|
||||
for mx in mx_hosts: |
|
||||
try: |
|
||||
smtp = smtplib.SMTP() |
|
||||
smtp.connect(mx[1]) |
|
||||
if not verify: |
|
||||
return True |
|
||||
status, _ = smtp.helo() |
|
||||
if status != 250: |
|
||||
continue |
|
||||
smtp.mail('') |
|
||||
status, _ = smtp.rcpt(email) |
|
||||
if status != 250: |
|
||||
return False |
|
||||
break |
|
||||
except smtplib.SMTPServerDisconnected: |
|
||||
# Server not permits verify user |
|
||||
break |
|
||||
except smtplib.SMTPConnectError: |
|
||||
continue |
|
||||
except (AssertionError, ServerError): |
|
||||
return False |
|
||||
return True |
|
||||
|
|
||||
# import sys |
|
||||
|
|
||||
# sys.modules[__name__], sys.modules['validate_email_module'] = validate_email, |
|
||||
# sys.modules[__name__] |
|
||||
# from validate_email_module import * |
|
||||
|
# © 2016 Camptocamp SA |
||||
|
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). |
||||
|
from openerp.addons.crm.validate_email import * # noqa |
@ -0,0 +1,2 @@ |
|||||
|
# TODO: change this to server-tools proper once this is merged |
||||
|
server-tools https://github.com/hbrunn/server-tools 8.0-base_manifest_extension |
Write
Preview
Loading…
Cancel
Save
Reference in new issue