You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
477 lines
18 KiB
477 lines
18 KiB
|
|
import base64
|
|
import logging
|
|
import os
|
|
import pstats
|
|
import re
|
|
import subprocess
|
|
import sys
|
|
from contextlib import contextmanager
|
|
from cProfile import Profile
|
|
|
|
import lxml.html
|
|
from psycopg2 import OperationalError, ProgrammingError
|
|
|
|
from odoo import _, api, exceptions, fields, models, sql_db, tools
|
|
|
|
if sys.version_info[0] >= 3:
|
|
from io import StringIO as IO
|
|
else:
|
|
from io import BytesIO as IO
|
|
|
|
DATETIME_FORMAT_FILE = "%Y%m%d_%H%M%S"
|
|
CPROFILE_EMPTY_CHARS = b"{0"
|
|
PGOPTIONS = {
|
|
'log_min_duration_statement': '0',
|
|
'client_min_messages': 'notice',
|
|
'log_min_messages': 'warning',
|
|
'log_min_error_statement': 'error',
|
|
'log_duration': 'off',
|
|
'log_error_verbosity': 'verbose',
|
|
'log_lock_waits': 'on',
|
|
'log_statement': 'none',
|
|
'log_temp_files': '0',
|
|
}
|
|
PGOPTIONS_ENV = ' '.join(["-c %s=%s" % (param, value)
|
|
for param, value in PGOPTIONS.items()])
|
|
PY_STATS_FIELDS = [
|
|
'ncalls',
|
|
'tottime', 'tt_percall',
|
|
'cumtime', 'ct_percall',
|
|
'file', 'lineno', 'method',
|
|
]
|
|
LINE_STATS_RE = re.compile(
|
|
r'(?P<%s>\d+/?\d+|\d+)\s+(?P<%s>\d+\.?\d+)\s+(?P<%s>\d+\.?\d+)\s+'
|
|
r'(?P<%s>\d+\.?\d+)\s+(?P<%s>\d+\.?\d+)\s+(?P<%s>.*):(?P<%s>\d+)'
|
|
r'\((?P<%s>.*)\)' % tuple(PY_STATS_FIELDS))
|
|
|
|
_logger = logging.getLogger(__name__)
|
|
|
|
|
|
class ProfilerProfilePythonLine(models.Model):
|
|
_name = 'profiler.profile.python.line'
|
|
_description = 'Profiler Python Line to save cProfiling results'
|
|
_rec_name = 'cprof_fname'
|
|
_order = 'cprof_cumtime DESC'
|
|
|
|
profile_id = fields.Many2one('profiler.profile', required=True,
|
|
ondelete='cascade')
|
|
cprof_tottime = fields.Float("Total time")
|
|
cprof_ncalls = fields.Float("Calls")
|
|
cprof_nrcalls = fields.Float("Recursive Calls")
|
|
cprof_ttpercall = fields.Float("Time per call")
|
|
cprof_cumtime = fields.Float("Cumulative time")
|
|
cprof_ctpercall = fields.Float("CT per call")
|
|
cprof_fname = fields.Char("Filename:lineno(method)")
|
|
|
|
|
|
class ProfilerProfile(models.Model):
|
|
_name = 'profiler.profile'
|
|
_description = 'Profiler Profile'
|
|
|
|
@api.model
|
|
def _find_loggers_path(self):
|
|
self.env.cr.execute("SHOW log_directory")
|
|
log_directory = self.env.cr.fetchone()[0]
|
|
self.env.cr.execute("SHOW log_filename")
|
|
log_filename = self.env.cr.fetchone()[0]
|
|
log_path = os.path.join(log_directory, log_filename)
|
|
if not os.path.isabs(log_path):
|
|
# It is relative path then join data_directory
|
|
self.env.cr.execute("SHOW data_directory")
|
|
data_dir = self.env.cr.fetchone()[0]
|
|
log_path = os.path.join(data_dir, log_path)
|
|
return log_path
|
|
|
|
name = fields.Char(required=True)
|
|
enable_python = fields.Boolean(default=True)
|
|
enable_postgresql = fields.Boolean(
|
|
default=False,
|
|
help="It requires postgresql server logs seudo-enabled")
|
|
use_py_index = fields.Boolean(
|
|
name="Get cProfiling report", default=False,
|
|
help="Index human-readable cProfile attachment."
|
|
"\nTo access this report, you must open the cprofile attachment view "
|
|
"using debug mode.\nWarning: Uses more resources.")
|
|
date_started = fields.Char(readonly=True)
|
|
date_finished = fields.Char(readonly=True)
|
|
state = fields.Selection([
|
|
('enabled', 'Enabled'),
|
|
('disabled', 'Disabled'),
|
|
], default='disabled', readonly=True, required=True)
|
|
description = fields.Text(readonly=True)
|
|
attachment_count = fields.Integer(compute="_compute_attachment_count")
|
|
pg_log_path = fields.Char(help="Getting the path to the logger",
|
|
default=_find_loggers_path)
|
|
pg_remote = fields.Char()
|
|
pg_stats_slowest_html = fields.Html(
|
|
"PostgreSQL Stats - Slowest", readonly=True)
|
|
pg_stats_time_consuming_html = fields.Html(
|
|
"PostgreSQL Stats - Time Consuming", readonly=True)
|
|
pg_stats_most_frequent_html = fields.Html(
|
|
"PostgreSQL Stats - Most Frequent", readonly=True)
|
|
py_stats_lines = fields.One2many(
|
|
"profiler.profile.python.line", "profile_id", "PY Stats Lines")
|
|
|
|
@api.multi
|
|
def _compute_attachment_count(self):
|
|
for record in self:
|
|
self.attachment_count = self.env['ir.attachment'].search_count([
|
|
('res_model', '=', self._name), ('res_id', '=', record.id)])
|
|
|
|
@api.onchange('enable_postgresql')
|
|
def onchange_enable_postgresql(self):
|
|
if not self.enable_postgresql:
|
|
return
|
|
self.env.cr.execute("SHOW config_file")
|
|
pg_config_file = self.env.cr.fetchone()[0]
|
|
db_host = tools.config.get('db_host')
|
|
if db_host == 'localhost' or db_host == '127.0.0.1':
|
|
db_host = False
|
|
if db_host:
|
|
pg_config_file = 'postgres@%s:%s' % (db_host, pg_config_file)
|
|
self.pg_remote = db_host
|
|
|
|
self.description = (
|
|
"You need seudo-enable logs from your "
|
|
"postgresql-server configuration file.\n\t- %s\n"
|
|
"or your can looking for the service using: "
|
|
"'ps aux | grep postgres'\n\n"
|
|
) % pg_config_file
|
|
self.description += """Adds the following parameters:
|
|
# Pre-enable logs
|
|
logging_collector=on
|
|
log_destination='stderr'
|
|
log_directory='/var/log/postgresql'
|
|
log_filename='postgresql.log'
|
|
log_rotation_age=0
|
|
log_checkpoints=on
|
|
log_hostname=on
|
|
log_line_prefix='%t [%p]: [%l-1] db=%d,user=%u '
|
|
log_connections=on
|
|
log_disconnections=on
|
|
lc_messages='en_US.UTF-8'
|
|
log_timezone='UTC'
|
|
|
|
Reload configuration using the following query:
|
|
- select pg_reload_conf()
|
|
Or restart the postgresql server service.
|
|
|
|
FYI This module will enable the following parameter from the client
|
|
It's not needed added them to configuration file if database user is
|
|
superuser or use PGOPTIONS environment variable in the terminal
|
|
that you start your odoo server.
|
|
If you don't add these parameters or PGOPTIONS this module will try do it.
|
|
# Enable logs from postgresql.conf
|
|
log_min_duration_statement=0
|
|
client_min_messages=notice
|
|
log_min_messages=warning
|
|
log_min_error_statement=error
|
|
log_duration=off
|
|
log_error_verbosity=verbose
|
|
log_lock_waits=on
|
|
log_statement=none
|
|
log_temp_files=0
|
|
|
|
# Or enable logs from PGOPTIONS environment variable before to start odoo
|
|
# server
|
|
export PGOPTIONS="-c log_min_duration_statement=0 \\
|
|
-c client_min_messages=notice -c log_min_messages=warning \\
|
|
-c log_min_error_statement=error -c log_connections=on \\
|
|
-c log_disconnections=on -c log_duration=off -c log_error_verbosity=verbose \\
|
|
-c log_lock_waits=on -c log_statement=none -c log_temp_files=0"
|
|
~/odoo_path/odoo-bin ...
|
|
"""
|
|
|
|
profile = Profile()
|
|
enabled = None
|
|
pglogs_enabled = None
|
|
|
|
# True to activate it False to inactivate None to do nothing
|
|
activate_deactivate_pglogs = None
|
|
|
|
# Params dict with values before to change it.
|
|
psql_params_original = {}
|
|
|
|
@api.model
|
|
def now_utc(self):
|
|
self.env.cr.execute("SHOW log_timezone")
|
|
zone = self.env.cr.fetchone()[0]
|
|
self.env.cr.execute("SELECT to_char(current_timestamp AT TIME "
|
|
"ZONE %s, 'YYYY-MM-DD HH24:MI:SS')", (zone,))
|
|
now = self.env.cr.fetchall()[0][0]
|
|
# now = fields.Datetime.to_string(
|
|
# fields.Datetime.context_timestamp(self, datetime.now()))
|
|
return now
|
|
|
|
@api.multi
|
|
def enable(self):
|
|
self.ensure_one()
|
|
if tools.config.get('workers'):
|
|
raise exceptions.UserError(
|
|
_("Start the odoo server using the parameter '--workers=0'"))
|
|
_logger.info("Enabling profiler")
|
|
self.write(dict(
|
|
date_started=self.now_utc(),
|
|
state='enabled'
|
|
))
|
|
ProfilerProfile.enabled = self.enable_python
|
|
self._reset_postgresql()
|
|
|
|
@api.multi
|
|
def _reset_postgresql(self):
|
|
if not self.enable_postgresql:
|
|
return
|
|
if ProfilerProfile.pglogs_enabled:
|
|
_logger.info("Using postgresql.conf or PGOPTIONS predefined.")
|
|
return
|
|
os.environ['PGOPTIONS'] = (
|
|
PGOPTIONS_ENV if self.state == 'enabled' else '')
|
|
self._reset_connection(self.state == 'enabled')
|
|
|
|
def _reset_connection(self, enable):
|
|
for connection in sql_db._Pool._connections:
|
|
with connection[0].cursor() as pool_cr:
|
|
params = (PGOPTIONS if enable
|
|
else ProfilerProfile.psql_params_original)
|
|
for param, value in params.items():
|
|
try:
|
|
pool_cr.execute('SET %s TO %s' % (param, value))
|
|
except (OperationalError, ProgrammingError) as oe:
|
|
pool_cr.connection.rollback()
|
|
raise exceptions.UserError(
|
|
_("It's not possible change parameter.\n%s\n"
|
|
"Please, disable postgresql or re-enable it "
|
|
"in order to read the instructions") % str(oe))
|
|
ProfilerProfile.activate_deactivate_pglogs = enable
|
|
|
|
def get_stats_string(self, cprofile_path):
|
|
pstats_stream = IO()
|
|
pstats_obj = pstats.Stats(cprofile_path, stream=pstats_stream)
|
|
pstats_obj.sort_stats('cumulative')
|
|
pstats_obj.print_stats()
|
|
pstats_stream.seek(0)
|
|
stats_string = pstats_stream.read()
|
|
pstats_stream = None
|
|
return stats_string
|
|
|
|
@api.multi
|
|
def dump_postgresql_logs(self, indexed=None):
|
|
self.ensure_one()
|
|
self.description = ''
|
|
pgbadger_cmd = self._get_pgbadger_command()
|
|
if pgbadger_cmd is None:
|
|
return
|
|
pgbadger_cmd_str = subprocess.list2cmdline(pgbadger_cmd)
|
|
self.description += (
|
|
'\nRunning the command: %s') % pgbadger_cmd_str
|
|
result = tools.exec_command_pipe(*pgbadger_cmd)
|
|
datas = result[1].read()
|
|
if not datas:
|
|
self.description += "\nPgbadger output is empty!"
|
|
return
|
|
fname = self._get_attachment_name("pg_stats", ".html")
|
|
self.env['ir.attachment'].create({
|
|
'name': fname,
|
|
'res_id': self.id,
|
|
'res_model': self._name,
|
|
'datas': base64.b64encode(datas),
|
|
'datas_fname': fname,
|
|
'description': 'pgbadger html output',
|
|
})
|
|
xpaths = [
|
|
'//*[@id="slowest-individual-queries"]',
|
|
'//*[@id="time-consuming-queries"]',
|
|
'//*[@id="most-frequent-queries"]',
|
|
]
|
|
# pylint: disable=unbalanced-tuple-unpacking
|
|
self.pg_stats_slowest_html, self.pg_stats_time_consuming_html, \
|
|
self.pg_stats_most_frequent_html = self._compute_pgbadger_html(
|
|
datas, xpaths)
|
|
|
|
@staticmethod
|
|
def _compute_pgbadger_html(html_doc, xpaths):
|
|
html = lxml.html.document_fromstring(html_doc)
|
|
result = []
|
|
for this_xpath in xpaths:
|
|
this_result = html.xpath(this_xpath)
|
|
result.append(
|
|
tools.html_sanitize(lxml.html.tostring(this_result[0])))
|
|
return result
|
|
|
|
@api.multi
|
|
def _get_pgbadger_command(self):
|
|
self.ensure_one()
|
|
# TODO: Catch early the following errors.
|
|
try:
|
|
pgbadger_bin = tools.find_in_path('pgbadger')
|
|
except IOError:
|
|
self.description += (
|
|
"\nInstall 'apt-get install pgbadger'")
|
|
try:
|
|
with open(self.pg_log_path, "r"):
|
|
pass
|
|
except IOError:
|
|
self.description += (
|
|
"\nCheck if exists and has permission to read the log file."
|
|
"\nMaybe running: chmod 604 '%s'"
|
|
) % self.pg_log_path
|
|
|
|
pgbadger_cmd = [
|
|
pgbadger_bin, '-f', 'stderr', '--sample', '15',
|
|
'-o', '-', '-x', 'html', '--quiet',
|
|
'-T', self.name,
|
|
'-d', self.env.cr.dbname,
|
|
'-b', self.date_started,
|
|
'-e', self.date_finished,
|
|
self.pg_log_path,
|
|
]
|
|
return pgbadger_cmd
|
|
|
|
def _get_attachment_name(self, prefix, suffix):
|
|
started = fields.Datetime.from_string(
|
|
self.date_started).strftime(DATETIME_FORMAT_FILE)
|
|
finished = fields.Datetime.from_string(
|
|
self.date_finished).strftime(DATETIME_FORMAT_FILE)
|
|
fname = '%s_%d_%s_to_%s%s' % (
|
|
prefix, self.id, started, finished, suffix)
|
|
return fname
|
|
|
|
@api.model
|
|
def dump_stats(self):
|
|
attachment = None
|
|
with tools.osutil.tempdir() as dump_dir:
|
|
cprofile_fname = self._get_attachment_name("py_stats", ".cprofile")
|
|
cprofile_path = os.path.join(dump_dir, cprofile_fname)
|
|
_logger.info("Dumping cProfile '%s'", cprofile_path)
|
|
ProfilerProfile.profile.dump_stats(cprofile_path)
|
|
with open(cprofile_path, "rb") as f_cprofile:
|
|
datas = f_cprofile.read()
|
|
if datas and datas != CPROFILE_EMPTY_CHARS:
|
|
attachment = self.env['ir.attachment'].create({
|
|
'name': cprofile_fname,
|
|
'res_id': self.id,
|
|
'res_model': self._name,
|
|
'datas': base64.b64encode(datas),
|
|
'datas_fname': cprofile_fname,
|
|
'description': 'cProfile dump stats',
|
|
})
|
|
_logger.info("A datas was saved, here %s", attachment.name)
|
|
try:
|
|
if self.use_py_index:
|
|
py_stats = self.get_stats_string(cprofile_path)
|
|
self.env['profiler.profile.python.line'].search([
|
|
('profile_id', '=', self.id)]).unlink()
|
|
for py_stat_line in py_stats.splitlines():
|
|
py_stat_line = py_stat_line.strip('\r\n ')
|
|
py_stat_line_match = LINE_STATS_RE.match(
|
|
py_stat_line) if py_stat_line else None
|
|
if not py_stat_line_match:
|
|
continue
|
|
data = dict((
|
|
field, py_stat_line_match.group(field))
|
|
for field in PY_STATS_FIELDS)
|
|
data['rcalls'], data['calls'] = (
|
|
"%(ncalls)s/%(ncalls)s" % data).split('/')[:2]
|
|
self.env['profiler.profile.python.line'].create({
|
|
'cprof_tottime': data['tottime'],
|
|
'cprof_ncalls': data['calls'],
|
|
'cprof_nrcalls': data['rcalls'],
|
|
'cprof_ttpercall': data['tt_percall'],
|
|
'cprof_cumtime': data['cumtime'],
|
|
'cprof_ctpercall': data['ct_percall'],
|
|
'cprof_fname': (
|
|
"%(file)s:%(lineno)s (%(method)s)" % data),
|
|
'profile_id': self.id,
|
|
})
|
|
attachment.index_content = py_stats
|
|
except IOError:
|
|
# Fancy feature but not stop process if fails
|
|
_logger.info("There was an error while getting the stats"
|
|
"from the cprofile_path")
|
|
# pylint: disable=unnecessary-pass
|
|
pass
|
|
self.dump_postgresql_logs()
|
|
_logger.info("cProfile stats stored.")
|
|
else:
|
|
_logger.info("cProfile stats empty.")
|
|
return attachment
|
|
|
|
@api.multi
|
|
def clear(self, reset_date=True):
|
|
self.ensure_one()
|
|
_logger.info("Clear profiler")
|
|
if reset_date:
|
|
self.date_started = self.now_utc()
|
|
ProfilerProfile.profile.clear()
|
|
|
|
@api.multi
|
|
def disable(self):
|
|
self.ensure_one()
|
|
_logger.info("Disabling profiler")
|
|
ProfilerProfile.enabled = False
|
|
self.state = 'disabled'
|
|
self.date_finished = self.now_utc()
|
|
self.dump_stats()
|
|
self.clear(reset_date=False)
|
|
self._reset_postgresql()
|
|
|
|
@staticmethod
|
|
@contextmanager
|
|
def profiling():
|
|
"""Thread local profile management, according to the shared "enabled"
|
|
"""
|
|
if ProfilerProfile.enabled:
|
|
_logger.debug("Catching profiling")
|
|
ProfilerProfile.profile.enable()
|
|
try:
|
|
yield
|
|
finally:
|
|
if ProfilerProfile.enabled:
|
|
ProfilerProfile.profile.disable()
|
|
|
|
@api.multi
|
|
def action_view_attachment(self):
|
|
attachments = self.env['ir.attachment'].search([
|
|
('res_model', '=', self._name), ('res_id', '=', self.id)])
|
|
action = self.env.ref("base.action_attachment").read()[0]
|
|
action['domain'] = [('id', 'in', attachments.ids)]
|
|
return action
|
|
|
|
@api.model
|
|
def set_pgoptions_enabled(self):
|
|
"""Verify if postgresql has configured the parameters for logging"""
|
|
ProfilerProfile.pglogs_enabled = True
|
|
pgoptions_enabled = bool(os.environ.get('PGOPTIONS'))
|
|
_logger.info('Logging enabled from environment '
|
|
'variable PGOPTIONS? %s', pgoptions_enabled)
|
|
if pgoptions_enabled:
|
|
return
|
|
pgparams_required = {
|
|
'log_min_duration_statement': '0',
|
|
}
|
|
for param, value in pgparams_required.items():
|
|
# pylint: disable=sql-injection
|
|
self.env.cr.execute("SHOW %s" % param)
|
|
db_value = self.env.cr.fetchone()[0].lower()
|
|
if value.lower() != db_value:
|
|
ProfilerProfile.pglogs_enabled = False
|
|
break
|
|
ProfilerProfile.psql_params_original = self.get_psql_params(
|
|
self.env.cr, PGOPTIONS.keys())
|
|
_logger.info('Logging enabled from postgresql.conf? %s',
|
|
ProfilerProfile.pglogs_enabled)
|
|
|
|
@staticmethod
|
|
def get_psql_params(cr, params):
|
|
result = {}
|
|
for param in set(params):
|
|
# pylint: disable=sql-injection
|
|
cr.execute('SHOW %s' % param)
|
|
result.update(cr.dictfetchone())
|
|
return result
|
|
|
|
@api.model
|
|
def _setup_complete(self):
|
|
self.set_pgoptions_enabled()
|
|
return super(ProfilerProfile, self)._setup_complete()
|