diff options
author | Lennart Weller <lhw@ring0.de> | 2017-01-24 15:21:09 +0000 |
---|---|---|
committer | Lennart Weller <lhw@ring0.de> | 2017-01-24 15:21:09 +0000 |
commit | 3ed3b02ed96ddab1c084811f3579b3a2aec83e04 (patch) | |
tree | 7a61ab288ae47800c4f11be5677d6ad8288dcd98 /python.d/varnish.chart.py | |
parent | New upstream version 1.4.0+dfsg (diff) | |
download | netdata-3ed3b02ed96ddab1c084811f3579b3a2aec83e04.tar.xz netdata-3ed3b02ed96ddab1c084811f3579b3a2aec83e04.zip |
New upstream version 1.5.0+dfsgupstream/1.5.0+dfsg
Diffstat (limited to '')
-rw-r--r-- | python.d/varnish.chart.py | 239 |
1 files changed, 239 insertions, 0 deletions
diff --git a/python.d/varnish.chart.py b/python.d/varnish.chart.py new file mode 100644 index 000000000..2b1512f4e --- /dev/null +++ b/python.d/varnish.chart.py @@ -0,0 +1,239 @@ +# -*- coding: utf-8 -*- +# Description: varnish netdata python.d module +# Author: l2isbad + +from base import SimpleService +from re import compile +from os import access as is_executable, X_OK +from subprocess import Popen, PIPE + + +# default module values (can be overridden per job in `config`) +# update_every = 2 +priority = 60000 +retries = 60 + +ORDER = ['session', 'hit_rate', 'chit_rate', 'expunge', 'threads', 'backend_health', 'memory_usage', 'bad', 'uptime'] + +CHARTS = {'backend_health': + {'lines': [['backend_conn', 'conn', 'incremental', 1, 1], + ['backend_unhealthy', 'unhealthy', 'incremental', 1, 1], + ['backend_busy', 'busy', 'incremental', 1, 1], + ['backend_fail', 'fail', 'incremental', 1, 1], + ['backend_reuse', 'reuse', 'incremental', 1, 1], + ['backend_recycle', 'resycle', 'incremental', 1, 1], + ['backend_toolate', 'toolate', 'incremental', 1, 1], + ['backend_retry', 'retry', 'incremental', 1, 1], + ['backend_req', 'req', 'incremental', 1, 1]], + 'options': [None, 'Backend health', 'connections', 'Backend health', 'varnish.backend_traf', 'line']}, + 'bad': + {'lines': [['sess_drop_b', None, 'incremental', 1, 1], + ['backend_unhealthy_b', None, 'incremental', 1, 1], + ['fetch_failed', None, 'incremental', 1, 1], + ['backend_busy_b', None, 'incremental', 1, 1], + ['threads_failed_b', None, 'incremental', 1, 1], + ['threads_limited_b', None, 'incremental', 1, 1], + ['threads_destroyed_b', None, 'incremental', 1, 1], + ['thread_queue_len_b', 'queue_len', 'absolute', 1, 1], + ['losthdr_b', None, 'incremental', 1, 1], + ['esi_errors_b', None, 'incremental', 1, 1], + ['esi_warnings_b', None, 'incremental', 1, 1], + ['sess_fail_b', None, 'incremental', 1, 1], + ['sc_pipe_overflow_b', None, 'incremental', 1, 1], + ['sess_pipe_overflow_b', None, 'incremental', 1, 1]], + 'options': [None, 'Misbehavior', 'problems', 'Problems summary', 'varnish.bad', 'line']}, + 'expunge': + {'lines': [['n_expired', 'expired', 'incremental', 1, 1], + ['n_lru_nuked', 'lru_nuked', 'incremental', 1, 1]], + 'options': [None, 'Object expunging', 'objects', 'Cache performance', 'varnish.expunge', 'line']}, + 'hit_rate': + {'lines': [['cache_hit_perc', 'hit', 'absolute', 1, 100], + ['cache_miss_perc', 'miss', 'absolute', 1, 100], + ['cache_hitpass_perc', 'hitpass', 'absolute', 1, 100]], + 'options': [None, 'All history hit rate ratio','percent', 'Cache performance', 'varnish.hit_rate', 'stacked']}, + 'chit_rate': + {'lines': [['cache_hit_cperc', 'hit', 'absolute', 1, 100], + ['cache_miss_cperc', 'miss', 'absolute', 1, 100], + ['cache_hitpass_cperc', 'hitpass', 'absolute', 1, 100]], + 'options': [None, 'Current poll hit rate ratio','percent', 'Cache performance', 'varnish.chit_rate', 'stacked']}, + 'memory_usage': + {'lines': [['s0.g_space', 'available', 'absolute', 1, 1048576], + ['s0.g_bytes', 'allocated', 'absolute', -1, 1048576]], + 'options': [None, 'Memory usage', 'megabytes', 'Memory usage', 'varnish.memory_usage', 'stacked']}, + 'session': + {'lines': [['sess_conn', 'sess_conn', 'incremental', 1, 1], + ['client_req', 'client_requests', 'incremental', 1, 1], + ['client_conn', 'client_conn', 'incremental', 1, 1], + ['client_drop', 'client_drop', 'incremental', 1, 1], + ['sess_dropped', 'sess_dropped', 'incremental', 1, 1]], + 'options': [None, 'Sessions', 'units', 'Client metrics', 'varnish.session', 'line']}, + 'threads': + {'lines': [['threads', None, 'absolute', 1, 1], + ['threads_created', 'created', 'incremental', 1, 1], + ['threads_failed', 'failed', 'incremental', 1, 1], + ['threads_limited', 'limited', 'incremental', 1, 1], + ['thread_queue_len', 'queue_len', 'incremental', 1, 1], + ['sess_queued', 'sess_queued', 'incremental', 1, 1]], + 'options': [None, 'Thread status', 'threads', 'Thread-related metrics', 'varnish.threads', 'line']}, + 'uptime': + {'lines': [['uptime', None, 'absolute', 1, 1]], + 'options': [None, 'Varnish uptime', 'seconds', 'Uptime', 'varnish.uptime', 'line']} +} + +DIRECTORIES = ['/bin/', '/usr/bin/', '/sbin/', '/usr/sbin/'] + + +class Service(SimpleService): + def __init__(self, configuration=None, name=None): + SimpleService.__init__(self, configuration=configuration, name=name) + try: + self.varnish = [''.join([directory, 'varnishstat']) for directory in DIRECTORIES + if is_executable(''.join([directory, 'varnishstat']), X_OK)][0] + except IndexError: + self.varnish = False + self.rgx_all = compile(r'([A-Z]+\.)?([\d\w_.]+)\s+(\d+)') + # Could be + # VBE.boot.super_backend.pipe_hdrbyte (new) + # or + # VBE.default2(127.0.0.2,,81).bereq_bodybytes (old) + # Regex result: [('super_backend', 'beresp_hdrbytes', '0'), ('super_backend', 'beresp_bodybytes', '0')] + self.rgx_bck = (compile(r'VBE.([\d\w_.]+)\(.*?\).(beresp[\w_]+)\s+(\d+)'), + compile(r'VBE\.[\d\w-]+\.([\w\d_]+).(beresp[\w_]+)\s+(\d+)')) + self.cache_prev = list() + + def check(self): + # Cant start without 'varnishstat' command + if not self.varnish: + self.error('\'varnishstat\' command was not found in %s or not executable by netdata' % DIRECTORIES) + return False + + # If command is present and we can execute it we need to make sure.. + # 1. STDOUT is not empty + reply = self._get_raw_data() + if not reply: + self.error('No output from \'varnishstat\' (not enough privileges?)') + return False + + # 2. Output is parsable (list is not empty after regex findall) + is_parsable = self.rgx_all.findall(reply) + if not is_parsable: + self.error('Cant parse output...') + return False + + # We need to find the right regex for backend parse + self.backend_list = self.rgx_bck[0].findall(reply)[::2] + if self.backend_list: + self.rgx_bck = self.rgx_bck[0] + else: + self.backend_list = self.rgx_bck[1].findall(reply)[::2] + self.rgx_bck = self.rgx_bck[1] + + # We are about to start! + self.create_charts() + + self.info('Plugin was started successfully') + return True + + def _get_raw_data(self): + try: + reply = Popen([self.varnish, '-1'], stdout=PIPE, stderr=PIPE, shell=False) + except OSError: + return None + + raw_data = reply.communicate()[0] + + if not raw_data: + return None + + return raw_data + + def _get_data(self): + """ + Format data received from shell command + :return: dict + """ + raw_data = self._get_raw_data() + data_all = self.rgx_all.findall(raw_data) + data_backend = self.rgx_bck.findall(raw_data) + + if not data_all: + return None + + # 1. ALL data from 'varnishstat -1'. t - type(MAIN, MEMPOOL etc) + to_netdata = {k: int(v) for t, k, v in data_all} + + # 2. ADD backend statistics + to_netdata.update({'_'.join([n, k]): int(v) for n, k, v in data_backend}) + + # 3. ADD additional keys to dict + # 3.1 Cache hit/miss/hitpass OVERALL in percent + cache_summary = sum([to_netdata.get('cache_hit', 0), to_netdata.get('cache_miss', 0), + to_netdata.get('cache_hitpass', 0)]) + to_netdata['cache_hit_perc'] = find_percent(to_netdata.get('cache_hit', 0), cache_summary, 10000) + to_netdata['cache_miss_perc'] = find_percent(to_netdata.get('cache_miss', 0), cache_summary, 10000) + to_netdata['cache_hitpass_perc'] = find_percent(to_netdata.get('cache_hitpass', 0), cache_summary, 10000) + + # 3.2 Cache hit/miss/hitpass CURRENT in percent + if self.cache_prev: + cache_summary = sum([to_netdata.get('cache_hit', 0), to_netdata.get('cache_miss', 0), + to_netdata.get('cache_hitpass', 0)]) - sum(self.cache_prev) + to_netdata['cache_hit_cperc'] = find_percent(to_netdata.get('cache_hit', 0) - self.cache_prev[0], cache_summary, 10000) + to_netdata['cache_miss_cperc'] = find_percent(to_netdata.get('cache_miss', 0) - self.cache_prev[1], cache_summary, 10000) + to_netdata['cache_hitpass_cperc'] = find_percent(to_netdata.get('cache_hitpass', 0) - self.cache_prev[2], cache_summary, 10000) + else: + to_netdata['cache_hit_cperc'] = 0 + to_netdata['cache_miss_cperc'] = 0 + to_netdata['cache_hitpass_cperc'] = 0 + + self.cache_prev = [to_netdata.get('cache_hit', 0), to_netdata.get('cache_miss', 0), to_netdata.get('cache_hitpass', 0)] + + # 3.3 Problems summary chart + for elem in ['backend_busy', 'backend_unhealthy', 'esi_errors', 'esi_warnings', 'losthdr', 'sess_drop', 'sc_pipe_overflow', + 'sess_fail', 'sess_pipe_overflow', 'threads_destroyed', 'threads_failed', 'threads_limited', 'thread_queue_len']: + if to_netdata.get(elem) is not None: + to_netdata[''.join([elem, '_b'])] = to_netdata.get(elem) + + # Ready steady go! + return to_netdata + + def create_charts(self): + # If 'all_charts' is true...ALL charts are displayed. If no only default + 'extra_charts' + #if self.configuration.get('all_charts'): + # self.order = EXTRA_ORDER + #else: + # try: + # extra_charts = list(filter(lambda chart: chart in EXTRA_ORDER, self.extra_charts.split())) + # except (AttributeError, NameError, ValueError): + # self.error('Extra charts disabled.') + # extra_charts = [] + + self.order = ORDER[:] + #self.order.extend(extra_charts) + + # Create static charts + #self.definitions = {chart: values for chart, values in CHARTS.items() if chart in self.order} + self.definitions = CHARTS + + # Create dynamic backend charts + if self.backend_list: + for backend in self.backend_list: + self.order.insert(0, ''.join([backend[0], '_resp_stats'])) + self.definitions.update({''.join([backend[0], '_resp_stats']): { + 'options': [None, + '%s response statistics' % backend[0].capitalize(), + "kilobit/s", + 'Backend response', + 'varnish.backend', + 'area'], + 'lines': [[''.join([backend[0], '_beresp_hdrbytes']), + 'header', 'incremental', 8, 1000], + [''.join([backend[0], '_beresp_bodybytes']), + 'body', 'incremental', -8, 1000]]}}) + + +def find_percent(value1, value2, multiply): + # If value2 is 0 return 0 + if not value2: + return 0 + else: + return round(float(value1) / float(value2) * multiply) |