Browse Source
Merge pull request #441 from hbrunn/8.0-base_partner_merge-cron
Merge pull request #441 from hbrunn/8.0-base_partner_merge-cron
[IMP] backport 9.0 mechanism8.0
Ronald Portier
6 years ago
committed by
GitHub
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
14 changed files with 244 additions and 1060 deletions
-
47base_partner_merge/README.rst
-
4base_partner_merge/__init__.py
-
15base_partner_merge/__openerp__.py
-
915base_partner_merge/base_partner_merge.py
-
29base_partner_merge/hooks.py
-
5base_partner_merge/models/__init__.py
-
16base_partner_merge/models/base_partner_merge.py
-
30base_partner_merge/models/base_partner_merge_automatic_wizard.py
-
21base_partner_merge/models/res_partner.py
-
3base_partner_merge/tests/__init__.py
-
79base_partner_merge/tests/test_merge.py
-
137base_partner_merge/validate_email.py
-
1base_partner_merge/views/base_partner_merge.xml
-
2oca_dependencies.txt
@ -1,2 +1,4 @@ |
|||
# -*- coding: utf-8 -*- |
|||
from . import base_partner_merge # NOQA |
|||
|
|||
from . import models |
|||
from .hooks import post_load_hook |
@ -1,17 +1,22 @@ |
|||
# -*- coding: utf-8 -*- |
|||
{ |
|||
'name': 'Base Partner Merge', |
|||
'author': "OpenERP S.A.,Odoo Community Association (OCA)", |
|||
'name': "Deduplicate Contacts (OCA)", |
|||
'author': "Camptocamp,Sunflower IT,Odoo Community Association (OCA)", |
|||
'category': 'Generic Modules/Base', |
|||
'version': '8.0.0.1.0', |
|||
'version': '8.0.1.0.0', |
|||
'license': 'AGPL-3', |
|||
'depends': [ |
|||
'base', |
|||
'mail' |
|||
'mail', |
|||
'base_manifest_extension', |
|||
], |
|||
'depends_if_installed': [ |
|||
'crm', |
|||
], |
|||
'data': [ |
|||
'security/ir.model.access.csv', |
|||
'base_partner_merge_view.xml', |
|||
'views/base_partner_merge.xml', |
|||
], |
|||
'installable': True, |
|||
'post_load': 'post_load_hook', |
|||
} |
@ -1,915 +0,0 @@ |
|||
# -*- coding: utf-8 -*- |
|||
|
|||
from __future__ import absolute_import |
|||
from email.utils import parseaddr |
|||
import functools |
|||
import htmlentitydefs |
|||
import itertools |
|||
import logging |
|||
import operator |
|||
import re |
|||
from ast import literal_eval |
|||
from openerp.tools import mute_logger |
|||
|
|||
# Validation Library https://pypi.python.org/pypi/validate_email/1.1 |
|||
from .validate_email import validate_email |
|||
|
|||
import openerp |
|||
import openerp.osv.fields as fields |
|||
from openerp.osv.orm import TransientModel, browse_record |
|||
from openerp.exceptions import except_orm |
|||
from openerp.tools.translate import _ |
|||
|
|||
pattern = re.compile(r"&(\w+?);") |
|||
|
|||
_logger = logging.getLogger('base.partner.merge') |
|||
|
|||
|
|||
# http://www.php2python.com/wiki/function.html-entity-decode/ |
|||
def html_entity_decode_char(m, defs=None): |
|||
if defs is None: |
|||
defs = htmlentitydefs.entitydefs |
|||
try: |
|||
return defs[m.group(1)] |
|||
except KeyError: |
|||
return m.group(0) |
|||
|
|||
|
|||
def html_entity_decode(string): |
|||
return pattern.sub(html_entity_decode_char, string) |
|||
|
|||
|
|||
def sanitize_email(partner_email): |
|||
assert isinstance(partner_email, basestring) and partner_email |
|||
|
|||
result = re.subn(r';|/|:', ',', |
|||
html_entity_decode(partner_email or ''))[0].split(',') |
|||
|
|||
emails = [parseaddr(email)[1] |
|||
for item in result |
|||
for email in item.split()] |
|||
|
|||
return [email.lower() |
|||
for email in emails |
|||
if validate_email(email)] |
|||
|
|||
|
|||
def is_integer_list(ids): |
|||
return all(isinstance(i, (int, long)) for i in ids) |
|||
|
|||
|
|||
class MergePartnerLine(TransientModel): |
|||
_name = 'base.partner.merge.line' |
|||
|
|||
_columns = { |
|||
'wizard_id': fields.many2one('base.partner.merge.automatic.wizard', |
|||
'Wizard'), |
|||
'min_id': fields.integer('MinID'), |
|||
'aggr_ids': fields.char('Ids', required=True), |
|||
} |
|||
|
|||
_order = 'min_id asc' |
|||
|
|||
|
|||
class MergePartnerAutomatic(TransientModel): |
|||
""" |
|||
The idea behind this wizard is to create a list of potential partners to |
|||
merge. We use two objects, the first one is the wizard for the end-user. |
|||
And the second will contain the partner list to merge. |
|||
|
|||
""" |
|||
_name = 'base.partner.merge.automatic.wizard' |
|||
|
|||
_columns = { |
|||
# Group by |
|||
'group_by_email': fields.boolean('Email'), |
|||
'group_by_name': fields.boolean('Name'), |
|||
'group_by_is_company': fields.boolean('Is Company'), |
|||
'group_by_vat': fields.boolean('VAT'), |
|||
'group_by_parent_id': fields.boolean('Parent Company'), |
|||
|
|||
'state': fields.selection([('option', 'Option'), |
|||
('selection', 'Selection'), |
|||
('finished', 'Finished')], |
|||
'State', |
|||
readonly=True, |
|||
required=True), |
|||
'number_group': fields.integer("Group of Contacts", readonly=True), |
|||
'current_line_id': fields.many2one('base.partner.merge.line', |
|||
'Current Line'), |
|||
'line_ids': fields.one2many('base.partner.merge.line', |
|||
'wizard_id', 'Lines'), |
|||
'partner_ids': fields.many2many('res.partner', string='Contacts'), |
|||
'dst_partner_id': fields.many2one('res.partner', |
|||
string='Destination Contact'), |
|||
|
|||
'exclude_contact': fields.boolean('A user associated to the contact'), |
|||
'exclude_journal_item': fields.boolean('Journal Items associated' |
|||
' to the contact'), |
|||
'maximum_group': fields.integer("Maximum of Group of Contacts"), |
|||
} |
|||
|
|||
def default_get(self, cr, uid, fields, context=None): |
|||
if context is None: |
|||
context = {} |
|||
res = super(MergePartnerAutomatic, self |
|||
).default_get(cr, uid, fields, context) |
|||
if (context.get('active_model') == 'res.partner' and |
|||
context.get('active_ids')): |
|||
partner_ids = context['active_ids'] |
|||
res['state'] = 'selection' |
|||
res['partner_ids'] = partner_ids |
|||
res['dst_partner_id'] = self._get_ordered_partner(cr, uid, |
|||
partner_ids, |
|||
context=context |
|||
)[-1].id |
|||
return res |
|||
|
|||
_defaults = { |
|||
'state': 'option' |
|||
} |
|||
|
|||
def get_fk_on(self, cr, table): |
|||
q = """ SELECT cl1.relname as table, |
|||
att1.attname as column |
|||
FROM pg_constraint as con, pg_class as cl1, pg_class as cl2, |
|||
pg_attribute as att1, pg_attribute as att2 |
|||
WHERE con.conrelid = cl1.oid |
|||
AND con.confrelid = cl2.oid |
|||
AND array_lower(con.conkey, 1) = 1 |
|||
AND con.conkey[1] = att1.attnum |
|||
AND att1.attrelid = cl1.oid |
|||
AND cl2.relname = %s |
|||
AND att2.attname = 'id' |
|||
AND array_lower(con.confkey, 1) = 1 |
|||
AND con.confkey[1] = att2.attnum |
|||
AND att2.attrelid = cl2.oid |
|||
AND con.contype = 'f' |
|||
""" |
|||
return cr.execute(q, (table,)) |
|||
|
|||
def _update_foreign_keys(self, cr, uid, src_partners, |
|||
dst_partner, context=None): |
|||
_logger.debug('_update_foreign_keys for dst_partner: %s for ' |
|||
'src_partners: %r', |
|||
dst_partner.id, |
|||
list(map(operator.attrgetter('id'), src_partners))) |
|||
|
|||
# find the many2one relation to a partner |
|||
proxy = self.pool.get('res.partner') |
|||
self.get_fk_on(cr, 'res_partner') |
|||
|
|||
# ignore two tables |
|||
|
|||
for table, column in cr.fetchall(): |
|||
if 'base_partner_merge_' in table: |
|||
continue |
|||
partner_ids = tuple(map(int, src_partners)) |
|||
|
|||
query = ("SELECT column_name FROM information_schema.columns" |
|||
" WHERE table_name LIKE '%s'") % (table) |
|||
cr.execute(query, ()) |
|||
columns = [] |
|||
for data in cr.fetchall(): |
|||
if data[0] != column: |
|||
columns.append(data[0]) |
|||
|
|||
query_dic = { |
|||
'table': table, |
|||
'column': column, |
|||
'value': columns[0], |
|||
} |
|||
if len(columns) <= 1: |
|||
# unique key treated |
|||
query = """ |
|||
UPDATE "%(table)s" as ___tu |
|||
SET %(column)s = %%s |
|||
WHERE |
|||
%(column)s = %%s AND |
|||
NOT EXISTS ( |
|||
SELECT 1 |
|||
FROM "%(table)s" as ___tw |
|||
WHERE |
|||
%(column)s = %%s AND |
|||
___tu.%(value)s = ___tw.%(value)s |
|||
)""" % query_dic |
|||
for partner_id in partner_ids: |
|||
cr.execute(query, (dst_partner.id, partner_id, |
|||
dst_partner.id)) |
|||
else: |
|||
cr.execute("SAVEPOINT recursive_partner_savepoint") |
|||
try: |
|||
query = ('UPDATE "%(table)s" SET %(column)s = %%s WHERE ' |
|||
'%(column)s IN %%s') % query_dic |
|||
cr.execute(query, (dst_partner.id, partner_ids,)) |
|||
|
|||
if (column == proxy._parent_name and |
|||
table == 'res_partner'): |
|||
query = """ |
|||
WITH RECURSIVE cycle(id, parent_id) AS ( |
|||
SELECT id, parent_id FROM res_partner |
|||
UNION |
|||
SELECT cycle.id, res_partner.parent_id |
|||
FROM res_partner, cycle |
|||
WHERE res_partner.id = cycle.parent_id |
|||
AND cycle.id != cycle.parent_id |
|||
) |
|||
SELECT id FROM cycle |
|||
WHERE id = parent_id AND id = %s |
|||
""" |
|||
cr.execute(query, (dst_partner.id,)) |
|||
if cr.fetchall(): |
|||
cr.execute("ROLLBACK TO SAVEPOINT " |
|||
"recursive_partner_savepoint") |
|||
finally: |
|||
cr.execute("RELEASE SAVEPOINT " |
|||
"recursive_partner_savepoint") |
|||
|
|||
def _update_reference_fields(self, cr, uid, src_partners, dst_partner, |
|||
context=None): |
|||
_logger.debug('_update_reference_fields for dst_partner: %s for ' |
|||
'src_partners: %r', |
|||
dst_partner.id, |
|||
list(map(operator.attrgetter('id'), src_partners))) |
|||
|
|||
def update_records(model, src, field_model='model', field_id='res_id', |
|||
context=None): |
|||
proxy = self.pool.get(model) |
|||
if proxy is None: |
|||
return |
|||
domain = [(field_model, '=', 'res.partner'), |
|||
(field_id, '=', src.id)] |
|||
ids = proxy.search(cr, openerp.SUPERUSER_ID, |
|||
domain, context=context) |
|||
if model == 'mail.followers': |
|||
# mail.followers have a set semantic |
|||
# unlink records that whould trigger a duplicate constraint |
|||
# on rewrite |
|||
src_objs = proxy.browse(cr, openerp.SUPERUSER_ID, |
|||
ids) |
|||
target_domain = [(field_model, '=', 'res.partner'), |
|||
(field_id, '=', dst_partner.id)] |
|||
target_ids = proxy.search(cr, openerp.SUPERUSER_ID, |
|||
target_domain, context=context) |
|||
dst_followers = proxy.browse(cr, openerp.SUPERUSER_ID, |
|||
target_ids).mapped('partner_id') |
|||
to_unlink = src_objs.filtered(lambda obj: |
|||
obj.partner_id in dst_followers) |
|||
to_rewrite = src_objs - to_unlink |
|||
to_unlink.unlink() |
|||
ids = to_rewrite.ids |
|||
|
|||
return proxy.write(cr, openerp.SUPERUSER_ID, ids, |
|||
{field_id: dst_partner.id}, context=context) |
|||
|
|||
update_records = functools.partial(update_records, context=context) |
|||
|
|||
for partner in src_partners: |
|||
update_records('base.calendar', src=partner, |
|||
field_model='model_id.model') |
|||
update_records('ir.attachment', src=partner, |
|||
field_model='res_model') |
|||
update_records('mail.followers', src=partner, |
|||
field_model='res_model') |
|||
update_records('mail.message', src=partner) |
|||
update_records('marketing.campaign.workitem', src=partner, |
|||
field_model='object_id.model') |
|||
update_records('ir.model.data', src=partner) |
|||
|
|||
proxy = self.pool['ir.model.fields'] |
|||
domain = [('ttype', '=', 'reference')] |
|||
record_ids = proxy.search(cr, openerp.SUPERUSER_ID, domain, |
|||
context=context) |
|||
|
|||
for record in proxy.browse(cr, openerp.SUPERUSER_ID, record_ids, |
|||
context=context): |
|||
try: |
|||
proxy_model = self.pool[record.model] |
|||
except KeyError: |
|||
# ignore old tables |
|||
continue |
|||
|
|||
if record.model == 'ir.property': |
|||
continue |
|||
|
|||
legacy = proxy_model._columns.get(record.name) |
|||
field_spec = proxy_model._fields.get(record.name) |
|||
|
|||
if not legacy or isinstance(legacy, fields.function) \ |
|||
or field_spec.compute: |
|||
continue |
|||
|
|||
for partner in src_partners: |
|||
domain = [ |
|||
(record.name, '=', 'res.partner,%d' % partner.id) |
|||
] |
|||
model_ids = proxy_model.search(cr, openerp.SUPERUSER_ID, |
|||
domain, context=context) |
|||
values = { |
|||
record.name: 'res.partner,%d' % dst_partner.id, |
|||
} |
|||
proxy_model.write(cr, openerp.SUPERUSER_ID, model_ids, values, |
|||
context=context) |
|||
|
|||
def _update_values(self, cr, uid, src_partners, dst_partner, context=None): |
|||
_logger.debug('_update_values for dst_partner: %s for src_partners: ' |
|||
'%r', |
|||
dst_partner.id, |
|||
list(map(operator.attrgetter('id'), src_partners))) |
|||
|
|||
columns = dst_partner._columns |
|||
|
|||
def write_serializer(column, item): |
|||
if isinstance(item, browse_record): |
|||
return item.id |
|||
else: |
|||
return item |
|||
|
|||
values = dict() |
|||
for column, field in columns.iteritems(): |
|||
if (field._type not in ('many2many', 'one2many') and |
|||
not isinstance(field, fields.function)): |
|||
for item in itertools.chain(src_partners, [dst_partner]): |
|||
if item[column]: |
|||
values[column] = write_serializer(column, |
|||
item[column]) |
|||
|
|||
values.pop('id', None) |
|||
parent_id = values.pop('parent_id', None) |
|||
dst_partner.write(values) |
|||
if parent_id and parent_id != dst_partner.id: |
|||
try: |
|||
dst_partner.write({'parent_id': parent_id}) |
|||
except except_orm: |
|||
_logger.info('Skip recursive partner hierarchies for ' |
|||
'parent_id %s of partner: %s', |
|||
parent_id, dst_partner.id) |
|||
|
|||
@mute_logger('openerp.osv.expression', 'openerp.osv.orm') |
|||
def _merge(self, cr, uid, partner_ids, dst_partner=None, context=None): |
|||
proxy = self.pool.get('res.partner') |
|||
|
|||
partner_ids = proxy.exists(cr, uid, list(partner_ids), |
|||
context=context) |
|||
if len(partner_ids) < 2: |
|||
return |
|||
|
|||
if len(partner_ids) > 3: |
|||
raise except_orm( |
|||
_('Error'), |
|||
_("For safety reasons, you cannot merge more than 3 contacts " |
|||
"together. You can re-open the wizard several times if " |
|||
"needed.")) |
|||
|
|||
if (openerp.SUPERUSER_ID != uid and |
|||
len(set(partner.email for partner |
|||
in proxy.browse(cr, uid, partner_ids, |
|||
context=context))) > 1): |
|||
raise except_orm( |
|||
_('Error'), |
|||
_("All contacts must have the same email. Only the " |
|||
"Administrator can merge contacts with different emails.")) |
|||
|
|||
if dst_partner and dst_partner.id in partner_ids: |
|||
src_partners = proxy.browse(cr, uid, |
|||
[id for id in partner_ids |
|||
if id != dst_partner.id], |
|||
context=context) |
|||
else: |
|||
ordered_partners = self._get_ordered_partner(cr, uid, partner_ids, |
|||
context) |
|||
dst_partner = ordered_partners[-1] |
|||
src_partners = ordered_partners[:-1] |
|||
_logger.info("dst_partner: %s", dst_partner.id) |
|||
|
|||
if (openerp.SUPERUSER_ID != uid and |
|||
self._model_is_installed( |
|||
cr, uid, 'account.move.line', context=context) and |
|||
self.pool['account.move.line'].search( |
|||
cr, openerp.SUPERUSER_ID, |
|||
[('partner_id', 'in', [partner.id for partner |
|||
in src_partners])], |
|||
context=context)): |
|||
raise except_orm( |
|||
_('Error'), |
|||
_("Only the destination contact may be linked to existing " |
|||
"Journal Items. Please ask the Administrator if you need to" |
|||
" merge several contacts linked to existing Journal " |
|||
"Items.")) |
|||
self._update_foreign_keys( |
|||
cr, uid, src_partners, dst_partner, context=context) |
|||
self._update_reference_fields( |
|||
cr, uid, src_partners, dst_partner, context=context) |
|||
self._update_values( |
|||
cr, uid, src_partners, dst_partner, context=context) |
|||
_logger.info('(uid = %s) merged the partners %r with %s', |
|||
uid, |
|||
list(map(operator.attrgetter('id'), src_partners)), |
|||
dst_partner.id) |
|||
dst_partner.message_post( |
|||
body='%s %s' % ( |
|||
_("Merged with the following partners:"), |
|||
", ".join( |
|||
'%s<%s>(ID %s)' % (p.name, p.email or 'n/a', p.id) |
|||
for p in src_partners |
|||
) |
|||
) |
|||
) |
|||
|
|||
for partner in src_partners: |
|||
partner.unlink() |
|||
|
|||
def clean_emails(self, cr, uid, context=None): |
|||
""" |
|||
Clean the email address of the partner, if there is an email field |
|||
with a minimum of two addresses, the system will create a new partner, |
|||
with the information of the previous one and will copy the new cleaned |
|||
email into the email field. |
|||
""" |
|||
if context is None: |
|||
context = {} |
|||
|
|||
proxy_model = self.pool['ir.model.fields'] |
|||
field_ids = proxy_model.search(cr, uid, |
|||
[('model', '=', 'res.partner'), |
|||
('ttype', 'like', '%2many')], |
|||
context=context) |
|||
fields = proxy_model.read(cr, uid, field_ids, context=context) |
|||
reset_fields = dict((field['name'], []) for field in fields) |
|||
|
|||
proxy_partner = self.pool['res.partner'] |
|||
context['active_test'] = False |
|||
ids = proxy_partner.search(cr, uid, [], context=context) |
|||
|
|||
fields = ['name', 'var' 'partner_id' 'is_company', 'email'] |
|||
partners = proxy_partner.read(cr, uid, ids, fields, context=context) |
|||
|
|||
partners.sort(key=operator.itemgetter('id')) |
|||
partners_len = len(partners) |
|||
|
|||
_logger.info('partner_len: %r', partners_len) |
|||
|
|||
for idx, partner in enumerate(partners): |
|||
if not partner['email']: |
|||
continue |
|||
|
|||
percent = (idx / float(partners_len)) * 100.0 |
|||
_logger.info('idx: %r', idx) |
|||
_logger.info('percent: %r', percent) |
|||
try: |
|||
emails = sanitize_email(partner['email']) |
|||
head, tail = emails[:1], emails[1:] |
|||
email = head[0] if head else False |
|||
|
|||
proxy_partner.write(cr, uid, [partner['id']], |
|||
{'email': email}, context=context) |
|||
|
|||
for email in tail: |
|||
values = dict(reset_fields, email=email) |
|||
proxy_partner.copy(cr, uid, partner['id'], values, |
|||
context=context) |
|||
|
|||
except Exception: |
|||
_logger.exception("There is a problem with this partner: %r", |
|||
partner) |
|||
raise |
|||
return True |
|||
|
|||
def close_cb(self, cr, uid, ids, context=None): |
|||
return {'type': 'ir.actions.act_window_close'} |
|||
|
|||
def _generate_query(self, fields, maximum_group=100): |
|||
group_fields = ', '.join(fields) |
|||
|
|||
filters = [] |
|||
for field in fields: |
|||
if field in ['email', 'name']: |
|||
filters.append((field, 'IS NOT', 'NULL')) |
|||
|
|||
criteria = ' AND '.join('%s %s %s' % (field, operator, value) |
|||
for field, operator, value in filters) |
|||
|
|||
text = [ |
|||
"SELECT min(id), array_agg(id)", |
|||
"FROM res_partner", |
|||
] |
|||
|
|||
if criteria: |
|||
text.append('WHERE %s' % criteria) |
|||
|
|||
text.extend([ |
|||
"GROUP BY %s" % group_fields, |
|||
"HAVING COUNT(*) >= 2", |
|||
"ORDER BY min(id)", |
|||
]) |
|||
|
|||
if maximum_group: |
|||
text.extend([ |
|||
"LIMIT %s" % maximum_group, |
|||
]) |
|||
|
|||
return ' '.join(text) |
|||
|
|||
def _compute_selected_groupby(self, this): |
|||
group_by_str = 'group_by_' |
|||
group_by_len = len(group_by_str) |
|||
|
|||
fields = [ |
|||
key[group_by_len:] |
|||
for key in self._columns.keys() |
|||
if key.startswith(group_by_str) |
|||
] |
|||
|
|||
groups = [ |
|||
field |
|||
for field in fields |
|||
if getattr(this, '%s%s' % (group_by_str, field), False) |
|||
] |
|||
|
|||
if not groups: |
|||
raise except_orm(_('Error'), |
|||
_("You have to specify a filter for your " |
|||
"selection")) |
|||
|
|||
return groups |
|||
|
|||
def next_cb(self, cr, uid, ids, context=None): |
|||
""" |
|||
Don't compute any thing |
|||
""" |
|||
context = dict(context or {}, active_test=False) |
|||
this = self.browse(cr, uid, ids[0], context=context) |
|||
if this.current_line_id: |
|||
this.current_line_id.unlink() |
|||
return self._next_screen(cr, uid, this, context) |
|||
|
|||
def _get_ordered_partner(self, cr, uid, partner_ids, context=None): |
|||
partners = self.pool.get('res.partner' |
|||
).browse(cr, uid, |
|||
list(partner_ids), |
|||
context=context) |
|||
ordered_partners = sorted( |
|||
sorted( |
|||
partners, |
|||
key=operator.attrgetter('create_date'), |
|||
reverse=True |
|||
), |
|||
key=operator.attrgetter('active'), |
|||
reverse=True |
|||
) |
|||
return ordered_partners |
|||
|
|||
def _next_screen(self, cr, uid, this, context=None): |
|||
this.refresh() |
|||
values = {} |
|||
if this.line_ids: |
|||
# in this case, we try to find the next record. |
|||
current_line = this.line_ids[0] |
|||
current_partner_ids = literal_eval(current_line.aggr_ids) |
|||
values.update({ |
|||
'current_line_id': current_line.id, |
|||
'partner_ids': [(6, 0, current_partner_ids)], |
|||
'dst_partner_id': self._get_ordered_partner( |
|||
cr, uid, |
|||
current_partner_ids, |
|||
context |
|||
)[-1].id, |
|||
'state': 'selection', |
|||
}) |
|||
else: |
|||
values.update({ |
|||
'current_line_id': False, |
|||
'partner_ids': [], |
|||
'state': 'finished', |
|||
}) |
|||
|
|||
this.write(values) |
|||
|
|||
return { |
|||
'type': 'ir.actions.act_window', |
|||
'res_model': this._name, |
|||
'res_id': this.id, |
|||
'view_mode': 'form', |
|||
'target': 'new', |
|||
} |
|||
|
|||
def _model_is_installed(self, cr, uid, model, context=None): |
|||
proxy = self.pool.get('ir.model') |
|||
domain = [('model', '=', model)] |
|||
return proxy.search_count(cr, uid, domain, context=context) > 0 |
|||
|
|||
def _partner_use_in(self, cr, uid, aggr_ids, models, context=None): |
|||
""" |
|||
Check if there is no occurence of this group of partner in the selected |
|||
model |
|||
""" |
|||
for model, field in models.iteritems(): |
|||
proxy = self.pool.get(model) |
|||
domain = [(field, 'in', aggr_ids)] |
|||
if proxy.search_count(cr, uid, domain, context=context): |
|||
return True |
|||
return False |
|||
|
|||
def compute_models(self, cr, uid, ids, context=None): |
|||
""" |
|||
Compute the different models needed by the system if you want to |
|||
exclude some partners. |
|||
""" |
|||
assert is_integer_list(ids) |
|||
|
|||
this = self.browse(cr, uid, ids[0], context=context) |
|||
|
|||
models = {} |
|||
if this.exclude_contact: |
|||
models['res.users'] = 'partner_id' |
|||
|
|||
if (self._model_is_installed( |
|||
cr, uid, 'account.move.line', context=context) and |
|||
this.exclude_journal_item): |
|||
models['account.move.line'] = 'partner_id' |
|||
|
|||
return models |
|||
|
|||
def _process_query(self, cr, uid, ids, query, context=None): |
|||
""" |
|||
Execute the select request and write the result in this wizard |
|||
""" |
|||
proxy = self.pool.get('base.partner.merge.line') |
|||
this = self.browse(cr, uid, ids[0], context=context) |
|||
models = self.compute_models(cr, uid, ids, context=context) |
|||
cr.execute(query) |
|||
|
|||
counter = 0 |
|||
for min_id, aggr_ids in cr.fetchall(): |
|||
if models and self._partner_use_in(cr, uid, aggr_ids, models, |
|||
context=context): |
|||
continue |
|||
values = { |
|||
'wizard_id': this.id, |
|||
'min_id': min_id, |
|||
'aggr_ids': aggr_ids, |
|||
} |
|||
|
|||
proxy.create(cr, uid, values, context=context) |
|||
counter += 1 |
|||
|
|||
values = { |
|||
'state': 'selection', |
|||
'number_group': counter, |
|||
} |
|||
|
|||
this.write(values) |
|||
|
|||
_logger.info("counter: %s", counter) |
|||
|
|||
def start_process_cb(self, cr, uid, ids, context=None): |
|||
""" |
|||
Start the process. |
|||
* Compute the selected groups (with duplication) |
|||
* If the user has selected the 'exclude_XXX' fields, avoid the |
|||
partners. |
|||
""" |
|||
assert is_integer_list(ids) |
|||
|
|||
context = dict(context or {}, active_test=False) |
|||
this = self.browse(cr, uid, ids[0], context=context) |
|||
groups = self._compute_selected_groupby(this) |
|||
query = self._generate_query(groups, this.maximum_group) |
|||
self._process_query(cr, uid, ids, query, context=context) |
|||
|
|||
return self._next_screen(cr, uid, this, context) |
|||
|
|||
def automatic_process_cb(self, cr, uid, ids, context=None): |
|||
assert is_integer_list(ids) |
|||
this = self.browse(cr, uid, ids[0], context=context) |
|||
this.start_process_cb() |
|||
this.refresh() |
|||
|
|||
for line in this.line_ids: |
|||
partner_ids = literal_eval(line.aggr_ids) |
|||
self._merge(cr, uid, partner_ids, context=context) |
|||
line.unlink() |
|||
cr.commit() |
|||
|
|||
this.write({'state': 'finished'}) |
|||
return { |
|||
'type': 'ir.actions.act_window', |
|||
'res_model': this._name, |
|||
'res_id': this.id, |
|||
'view_mode': 'form', |
|||
'target': 'new', |
|||
} |
|||
|
|||
def parent_migration_process_cb(self, cr, uid, ids, context=None): |
|||
assert is_integer_list(ids) |
|||
|
|||
context = dict(context or {}, active_test=False) |
|||
this = self.browse(cr, uid, ids[0], context=context) |
|||
|
|||
query = """ |
|||
SELECT |
|||
min(p1.id), |
|||
array_agg(DISTINCT p1.id) |
|||
FROM |
|||
res_partner as p1 |
|||
INNER join |
|||
res_partner as p2 |
|||
ON |
|||
p1.email = p2.email AND |
|||
p1.name = p2.name AND |
|||
(p1.parent_id = p2.id OR p1.id = p2.parent_id) |
|||
WHERE |
|||
p2.id IS NOT NULL |
|||
GROUP BY |
|||
p1.email, |
|||
p1.name, |
|||
CASE WHEN p1.parent_id = p2.id THEN p2.id |
|||
ELSE p1.id |
|||
END |
|||
HAVING COUNT(*) >= 2 |
|||
ORDER BY |
|||
min(p1.id) |
|||
""" |
|||
|
|||
self._process_query(cr, uid, ids, query, context=context) |
|||
|
|||
for line in this.line_ids: |
|||
partner_ids = literal_eval(line.aggr_ids) |
|||
self._merge(cr, uid, partner_ids, context=context) |
|||
line.unlink() |
|||
cr.commit() |
|||
|
|||
this.write({'state': 'finished'}) |
|||
|
|||
cr.execute(""" |
|||
UPDATE |
|||
res_partner |
|||
SET |
|||
is_company = NULL, |
|||
parent_id = NULL |
|||
WHERE |
|||
parent_id = id |
|||
""") |
|||
|
|||
return { |
|||
'type': 'ir.actions.act_window', |
|||
'res_model': this._name, |
|||
'res_id': this.id, |
|||
'view_mode': 'form', |
|||
'target': 'new', |
|||
} |
|||
|
|||
def update_all_process_cb(self, cr, uid, ids, context=None): |
|||
assert is_integer_list(ids) |
|||
|
|||
# WITH RECURSIVE cycle(id, parent_id) AS ( |
|||
# SELECT id, parent_id FROM res_partner |
|||
# UNION |
|||
# SELECT cycle.id, res_partner.parent_id |
|||
# FROM res_partner, cycle |
|||
# WHERE res_partner.id = cycle.parent_id AND |
|||
# cycle.id != cycle.parent_id |
|||
# ) |
|||
# UPDATE res_partner |
|||
# SET parent_id = NULL |
|||
# WHERE id in (SELECT id FROM cycle WHERE id = parent_id); |
|||
|
|||
this = self.browse(cr, uid, ids[0], context=context) |
|||
|
|||
self.parent_migration_process_cb(cr, uid, ids, context=None) |
|||
|
|||
list_merge = [ |
|||
{'group_by_vat': True, |
|||
'group_by_email': True, |
|||
'group_by_name': True}, |
|||
# {'group_by_name': True, |
|||
# 'group_by_is_company': True, |
|||
# 'group_by_parent_id': True}, |
|||
# {'group_by_email': True, |
|||
# 'group_by_is_company': True, |
|||
# 'group_by_parent_id': True}, |
|||
# {'group_by_name': True, |
|||
# 'group_by_vat': True, |
|||
# 'group_by_is_company': True, |
|||
# 'exclude_journal_item': True}, |
|||
# {'group_by_email': True, |
|||
# 'group_by_vat': True, |
|||
# 'group_by_is_company': True, |
|||
# 'exclude_journal_item': True}, |
|||
# {'group_by_email': True, |
|||
# 'group_by_is_company': True, |
|||
# 'exclude_contact': True, |
|||
# 'exclude_journal_item': True}, |
|||
# {'group_by_name': True, |
|||
# 'group_by_is_company': True, |
|||
# 'exclude_contact': True, |
|||
# 'exclude_journal_item': True} |
|||
] |
|||
|
|||
for merge_value in list_merge: |
|||
id = self.create(cr, uid, merge_value, context=context) |
|||
self.automatic_process_cb(cr, uid, [id], context=context) |
|||
|
|||
cr.execute(""" |
|||
UPDATE |
|||
res_partner |
|||
SET |
|||
is_company = NULL |
|||
WHERE |
|||
parent_id IS NOT NULL AND |
|||
is_company IS NOT NULL |
|||
""") |
|||
|
|||
# cr.execute(""" |
|||
# UPDATE |
|||
# res_partner as p1 |
|||
# SET |
|||
# is_company = NULL, |
|||
# parent_id = ( |
|||
# SELECT p2.id |
|||
# FROM res_partner as p2 |
|||
# WHERE p2.email = p1.email AND |
|||
# p2.parent_id != p2.id |
|||
# LIMIT 1 |
|||
# ) |
|||
# WHERE |
|||
# p1.parent_id = p1.id |
|||
# """) |
|||
|
|||
return self._next_screen(cr, uid, this, context) |
|||
|
|||
def merge_cb(self, cr, uid, ids, context=None): |
|||
assert is_integer_list(ids) |
|||
|
|||
context = dict(context or {}, active_test=False) |
|||
this = self.browse(cr, uid, ids[0], context=context) |
|||
|
|||
partner_ids = set(map(int, this.partner_ids)) |
|||
if not partner_ids: |
|||
this.write({'state': 'finished'}) |
|||
return { |
|||
'type': 'ir.actions.act_window', |
|||
'res_model': this._name, |
|||
'res_id': this.id, |
|||
'view_mode': 'form', |
|||
'target': 'new', |
|||
} |
|||
|
|||
self._merge(cr, uid, partner_ids, this.dst_partner_id, |
|||
context=context) |
|||
|
|||
if this.current_line_id: |
|||
this.current_line_id.unlink() |
|||
|
|||
return self._next_screen(cr, uid, this, context) |
|||
|
|||
def auto_set_parent_id(self, cr, uid, ids, context=None): |
|||
assert is_integer_list(ids) |
|||
|
|||
# select partner who have one least invoice |
|||
partner_treated = ['@gmail.com'] |
|||
cr.execute(""" SELECT p.id, p.email |
|||
FROM res_partner as p |
|||
LEFT JOIN account_invoice as a |
|||
ON p.id = a.partner_id AND a.state in ('open','paid') |
|||
WHERE p.grade_id is NOT NULL |
|||
GROUP BY p.id |
|||
ORDER BY COUNT(a.id) DESC |
|||
""") |
|||
re_email = re.compile(r".*@") |
|||
for id, email in cr.fetchall(): |
|||
# check email domain |
|||
email = re_email.sub("@", email or "") |
|||
if not email or email in partner_treated: |
|||
continue |
|||
partner_treated.append(email) |
|||
|
|||
# don't update the partners if they are more of one who have |
|||
# invoice |
|||
cr.execute(""" |
|||
SELECT * |
|||
FROM res_partner as p |
|||
WHERE p.id != %s AND p.email LIKE '%%%s' AND |
|||
EXISTS (SELECT * FROM account_invoice as a |
|||
WHERE p.id = a.partner_id |
|||
AND a.state in ('open','paid')) |
|||
""" % (id, email)) |
|||
|
|||
if len(cr.fetchall()) > 1: |
|||
_logger.info("%s MORE OF ONE COMPANY", email) |
|||
continue |
|||
|
|||
# to display changed values |
|||
cr.execute(""" SELECT id,email |
|||
FROM res_partner |
|||
WHERE parent_id != %s |
|||
AND id != %s AND email LIKE '%%%s' |
|||
""" % (id, id, email)) |
|||
_logger.info("%r", cr.fetchall()) |
|||
|
|||
# upgrade |
|||
cr.execute(""" UPDATE res_partner |
|||
SET parent_id = %s |
|||
WHERE id != %s AND email LIKE '%%%s' |
|||
""" % (id, id, email)) |
|||
return False |
@ -0,0 +1,29 @@ |
|||
# -*- coding: utf-8 -*- |
|||
# © 2017 Therp BV <http://therp.nl> |
|||
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html). |
|||
import inspect |
|||
from openerp.models import MetaModel |
|||
|
|||
|
|||
def post_load_hook(): |
|||
""" |
|||
We try to be smart here: If the crm module is to be loaded too |
|||
(or is already loaded), we remove our own models again in order not to |
|||
clash with the CRM ones: https://github.com/OCA/partner-contact/issues/283 |
|||
""" |
|||
for frame, filename, lineno, funcname, line, index in inspect.stack(): |
|||
# walk up the stack until we're in load_module_graph |
|||
if 'graph' in frame.f_locals: |
|||
graph = frame.f_locals['graph'] |
|||
package = frame.f_locals['package'] |
|||
if any(p.name == 'crm' for p in graph): |
|||
# so crm is installed, then we need to remove the models |
|||
# we pull from CRM again |
|||
MetaModel.module_to_models['base_partner_merge'] = [ |
|||
cls |
|||
for cls in MetaModel.module_to_models['base_partner_merge'] |
|||
if 'NoCRM' not in cls.__name__ |
|||
] |
|||
# and in this case, we also don't want to load our xml file |
|||
package.data['data'].remove('views/base_partner_merge.xml') |
|||
break |
@ -0,0 +1,5 @@ |
|||
# -*- coding: utf-8 -*- |
|||
|
|||
from . import base_partner_merge # noqa |
|||
from . import res_partner |
|||
from . import base_partner_merge_automatic_wizard |
@ -0,0 +1,16 @@ |
|||
# -*- coding: utf-8 -*- |
|||
# © 2016 Camptocamp SA |
|||
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). |
|||
from openerp.addons.crm.base_partner_merge import * # noqa |
|||
|
|||
|
|||
class NoCRMResPartner(ResPartner): # noqa |
|||
_module = 'base_partner_merge' |
|||
|
|||
|
|||
class NoCRMMergePartnerLine(MergePartnerLine): # noqa |
|||
_module = 'base_partner_merge' |
|||
|
|||
|
|||
class NoCRMMergePartnerAutomatic(MergePartnerAutomatic): # noqa |
|||
_module = 'base_partner_merge' |
@ -0,0 +1,30 @@ |
|||
# -*- coding: utf-8 -*- |
|||
# © 2017 Sunflower IT <http://sunflowerweb.nl> |
|||
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html). |
|||
|
|||
from openerp import fields, models, api |
|||
from ast import literal_eval |
|||
|
|||
|
|||
class MergePartnerAutomatic(models.TransientModel): |
|||
_inherit = 'base.partner.merge.automatic.wizard' |
|||
|
|||
# Enable deduplicating by reference |
|||
group_by_ref = fields.Boolean('Reference') |
|||
|
|||
@api.multi |
|||
def _process_query(self, query): |
|||
ret = super(MergePartnerAutomatic, self)._process_query(query) |
|||
|
|||
# If 'extra_domain', deduplicate only the records matching the domain |
|||
extra_domain = self.env.context.get('partner_merge_domain', []) |
|||
if extra_domain: |
|||
for line in self.line_ids: |
|||
domain = [('id', 'in', literal_eval(line.aggr_ids))] |
|||
domain.extend(extra_domain) |
|||
aggr_ids = self.env['res.partner'].search(domain).ids |
|||
if len(aggr_ids) > 1: |
|||
line.aggr_ids = str(aggr_ids) |
|||
else: |
|||
line.unlink() |
|||
return ret |
@ -0,0 +1,21 @@ |
|||
# -*- coding: utf-8 -*- |
|||
# © 2017 Sunflower IT <http://sunflowerweb.nl> |
|||
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html). |
|||
|
|||
from openerp import models, api |
|||
|
|||
|
|||
class ResPartnerChanges(models.Model): |
|||
_inherit = 'res.partner' |
|||
|
|||
@api.model |
|||
def deduplicate_on_fields(self, fields_list, domain=None): |
|||
""" Merge contacts """ |
|||
wizard_obj = self.env['base.partner.merge.automatic.wizard'] |
|||
if domain: |
|||
wizard_obj = wizard_obj.with_context(partner_merge_domain=domain) |
|||
params = {'state': 'option'} |
|||
for field in fields_list: |
|||
params['group_by_%s' % (field,)] = True |
|||
wizard = wizard_obj.create(params) |
|||
wizard.automatic_process_cb() |
@ -0,0 +1,3 @@ |
|||
# -*- coding: utf-8 -*- |
|||
|
|||
from . import test_merge |
@ -0,0 +1,79 @@ |
|||
# -*- coding: utf-8 -*- |
|||
# © 2017 Sunflower IT <http://sunflowerweb.nl> |
|||
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html). |
|||
|
|||
from openerp.tests.common import TransactionCase |
|||
|
|||
|
|||
class PartnerMergeTestCase(TransactionCase): |
|||
"""Tests for Partner Merge""" |
|||
|
|||
def setUp(self): |
|||
super(PartnerMergeTestCase, self).setUp() |
|||
self.partner = self.env['res.partner'] |
|||
self.merge_wizard = \ |
|||
self.env['base.partner.merge.automatic.wizard'] |
|||
self.donald_domain = [('name', '=', 'Donald Duck')] |
|||
self.mickey_domain = [('name', '=', 'Mickey Mouse')] |
|||
|
|||
def _unlink_all(self): |
|||
self.partner.search(self.donald_domain).unlink() |
|||
self.partner.search(self.mickey_domain).unlink() |
|||
|
|||
def _count_donalds_mickeys(self, donalds, mickeys): |
|||
self.assertEquals( |
|||
len(self.partner.search(self.donald_domain)), donalds) |
|||
self.assertEquals( |
|||
len(self.partner.search(self.mickey_domain)), mickeys) |
|||
|
|||
def _create_duplicates(self, field1, value1, field2, values2): |
|||
for value2 in values2: |
|||
self.partner.create({ |
|||
field1: value1, |
|||
field2: value2, |
|||
}) |
|||
|
|||
def test_10_name_merge(self): |
|||
""" Merge users with duplicate names """ |
|||
self._unlink_all() |
|||
self._create_duplicates( |
|||
'name', 'Donald Duck', |
|||
'email', ['donald@therp.nl', 'donald@sunflowerweb.nl'], |
|||
) |
|||
self._create_duplicates( |
|||
'name', 'Mickey Mouse', |
|||
'email', ['mickey@therp.nl', 'mickey@sunflowerweb.nl'], |
|||
) |
|||
self._count_donalds_mickeys(2, 2) |
|||
# Merge all names that start with 'D', |
|||
self.partner.deduplicate_on_fields( |
|||
['name'], domain=[('name', 'like', 'D%')], |
|||
) |
|||
self._count_donalds_mickeys(1, 2) |
|||
|
|||
def test_20_ref_merge(self): |
|||
""" Merge users with duplicate references """ |
|||
self._unlink_all() |
|||
self._create_duplicates( |
|||
'ref', 'DD123', |
|||
'name', ['Donald Duck', 'Mickey Mouse'], |
|||
) |
|||
# Merge on reference, leaving out guys that have no ref |
|||
self.partner.deduplicate_on_fields( |
|||
['ref'], domain=[('ref', '=', 'DD123')], |
|||
) |
|||
# Test if only one remains after |
|||
partners = self.partner.search([('ref', '=', 'DD123')]) |
|||
self.assertEquals(len(partners), 1) |
|||
|
|||
def test_30_ref_merge(self): |
|||
""" Fringe case: three guys, two to merge """ |
|||
self._unlink_all() |
|||
self._create_duplicates( |
|||
'ref', 'DD123', |
|||
'name', ['Donald Duck', 'Donald Duck', 'Mickey Mouse'], |
|||
) |
|||
self.partner.deduplicate_on_fields( |
|||
['ref'], domain=[('name', '=', 'Donald Duck')], |
|||
) |
|||
self._count_donalds_mickeys(1, 1) |
@ -1,135 +1,4 @@ |
|||
# -*- coding: utf-8 -*- |
|||
# RFC 2822 - style email validation for Python |
|||
# (c) 2012 Syrus Akbary <me@syrusakbary.com> |
|||
# Extended from (c) 2011 Noel Bush <noel@aitools.org> |
|||
# for support of mx and user check |
|||
# This code is made available to you under the GNU LGPL v3. |
|||
# |
|||
# This module provides a single method, valid_email_address(), |
|||
# which returns True or False to indicate whether a given address |
|||
# is valid according to the 'addr-spec' part of the specification |
|||
# given in RFC 2822. Ideally, we would like to find this |
|||
# in some other library, already thoroughly tested and well- |
|||
# maintained. The standard Python library email.utils |
|||
# contains a parse_addr() function, but it is not sufficient |
|||
# to detect many malformed addresses. |
|||
# |
|||
# This implementation aims to be faithful to the RFC, with the |
|||
# exception of a circular definition (see comments below), and |
|||
# with the omission of the pattern components marked as "obsolete". |
|||
|
|||
import re |
|||
import smtplib |
|||
|
|||
try: |
|||
import DNS |
|||
ServerError = DNS.ServerError |
|||
except: |
|||
DNS = None |
|||
|
|||
class ServerError(Exception): |
|||
pass |
|||
# All we are really doing is comparing the input string to one |
|||
# gigantic regular expression. But building that regexp, and |
|||
# ensuring its correctness, is made much easier by assembling it |
|||
# from the "tokens" defined by the RFC. Each of these tokens is |
|||
# tested in the accompanying unit test file. |
|||
# |
|||
# The section of RFC 2822 from which each pattern component is |
|||
# derived is given in an accompanying comment. |
|||
# |
|||
# (To make things simple, every string below is given as 'raw', |
|||
# even when it's not strictly necessary. This way we don't forget |
|||
# when it is necessary.) |
|||
# |
|||
WSP = r'[ \t]' # see 2.2.2. Structured Header Field Bodies |
|||
CRLF = r'(?:\r\n)' # see 2.2.3. Long Header Fields |
|||
NO_WS_CTL = r'\x01-\x08\x0b\x0c\x0f-\x1f\x7f' # see 3.2.1. Primitive Tokens |
|||
QUOTED_PAIR = r'(?:\\.)' # see 3.2.2. Quoted characters |
|||
FWS = r'(?:(?:{0}*{1})?{0}+)'.format(WSP, CRLF) |
|||
# see 3.2.3. Folding white space and comments |
|||
CTEXT = r'[{0}\x21-\x27\x2a-\x5b\x5d-\x7e]'.format( |
|||
NO_WS_CTL) # see 3.2.3 |
|||
# see 3.2.3 (NB: The RFC includes COMMENT here as well, but that would be |
|||
# circular.) |
|||
CCONTENT = r'(?:{0}|{1})'.format(CTEXT, QUOTED_PAIR) |
|||
COMMENT = r'\((?:{0}?{1})*{0}?\)'.format( |
|||
FWS, CCONTENT) # see 3.2.3 |
|||
CFWS = r'(?:{0}?{1})*(?:{0}?{1}|{0})'.format( |
|||
FWS, COMMENT) # see 3.2.3 |
|||
ATEXT = r'[\w!#$%&\'\*\+\-/=\?\^`\{\|\}~]' # see 3.2.4. Atom |
|||
ATOM = r'{0}?{1}+{0}?'.format(CFWS, ATEXT) |
|||
# see 3.2.4 |
|||
DOT_ATOM_TEXT = r'{0}+(?:\.{0}+)*'.format( |
|||
ATEXT) # see 3.2.4 |
|||
DOT_ATOM = r'{0}?{1}{0}?'.format( |
|||
CFWS, DOT_ATOM_TEXT) # see 3.2.4 |
|||
QTEXT = r'[{0}\x21\x23-\x5b\x5d-\x7e]'.format( |
|||
NO_WS_CTL) # see 3.2.5. Quoted strings |
|||
QCONTENT = r'(?:{0}|{1})'.format(QTEXT, QUOTED_PAIR) |
|||
# see 3.2.5 |
|||
QUOTED_STRING = r'{0}?"(?:{1}?{2})*{1}?"{0}?'.format(CFWS, FWS, QCONTENT) |
|||
LOCAL_PART = r'(?:{0}|{1})'.format(DOT_ATOM, QUOTED_STRING) |
|||
# see 3.4.1. Addr-spec specification |
|||
DTEXT = r'[{0}\x21-\x5a\x5e-\x7e]'.format( |
|||
NO_WS_CTL) # see 3.4.1 |
|||
DCONTENT = r'(?:{0}|{1})'.format(DTEXT, QUOTED_PAIR) |
|||
# see 3.4.1 |
|||
DOMAIN_LITERAL = r'{0}?\[(?:{1}?{2})*{1}?\]{0}?'.format( |
|||
CFWS, FWS, DCONTENT) # see 3.4.1 |
|||
DOMAIN = r'(?:{0}|{1})'.format(DOT_ATOM, DOMAIN_LITERAL) |
|||
# see 3.4.1 |
|||
ADDR_SPEC = r'{0}@{1}'.format( |
|||
LOCAL_PART, DOMAIN) # see 3.4.1 |
|||
|
|||
# A valid address will match exactly the 3.4.1 addr-spec. |
|||
VALID_ADDRESS_REGEXP = '^' + ADDR_SPEC + '$' |
|||
|
|||
|
|||
def validate_email(email, check_mx=False, verify=False): |
|||
"""Indicate whether the given string is a valid email address |
|||
according to the 'addr-spec' portion of RFC 2822 (see section |
|||
3.4.1). Parts of the spec that are marked obsolete are *not* |
|||
included in this test, and certain arcane constructions that |
|||
depend on circular definitions in the spec may not pass, but in |
|||
general this should correctly identify any email address likely |
|||
to be in use as of 2011.""" |
|||
try: |
|||
assert re.match(VALID_ADDRESS_REGEXP, email) is not None |
|||
check_mx |= verify |
|||
if check_mx: |
|||
if not DNS: |
|||
raise Exception('For check the mx records or check if the ' |
|||
'email exists you must have installed pyDNS ' |
|||
'python package') |
|||
DNS.DiscoverNameServers() |
|||
hostname = email[email.find('@') + 1:] |
|||
mx_hosts = DNS.mxlookup(hostname) |
|||
for mx in mx_hosts: |
|||
try: |
|||
smtp = smtplib.SMTP() |
|||
smtp.connect(mx[1]) |
|||
if not verify: |
|||
return True |
|||
status, _ = smtp.helo() |
|||
if status != 250: |
|||
continue |
|||
smtp.mail('') |
|||
status, _ = smtp.rcpt(email) |
|||
if status != 250: |
|||
return False |
|||
break |
|||
except smtplib.SMTPServerDisconnected: |
|||
# Server not permits verify user |
|||
break |
|||
except smtplib.SMTPConnectError: |
|||
continue |
|||
except (AssertionError, ServerError): |
|||
return False |
|||
return True |
|||
|
|||
# import sys |
|||
|
|||
# sys.modules[__name__], sys.modules['validate_email_module'] = validate_email, |
|||
# sys.modules[__name__] |
|||
# from validate_email_module import * |
|||
# © 2016 Camptocamp SA |
|||
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). |
|||
from openerp.addons.crm.validate_email import * # noqa |
@ -0,0 +1,2 @@ |
|||
# TODO: change this to server-tools proper once this is merged |
|||
server-tools https://github.com/hbrunn/server-tools 8.0-base_manifest_extension |
Write
Preview
Loading…
Cancel
Save
Reference in new issue