summaryrefslogtreecommitdiffstats
path: root/python.d/varnish.chart.py
diff options
context:
space:
mode:
authorLennart Weller <lhw@ring0.de>2017-01-24 15:21:09 +0000
committerLennart Weller <lhw@ring0.de>2017-01-24 15:21:09 +0000
commit3ed3b02ed96ddab1c084811f3579b3a2aec83e04 (patch)
tree7a61ab288ae47800c4f11be5677d6ad8288dcd98 /python.d/varnish.chart.py
parentNew upstream version 1.4.0+dfsg (diff)
downloadnetdata-3ed3b02ed96ddab1c084811f3579b3a2aec83e04.tar.xz
netdata-3ed3b02ed96ddab1c084811f3579b3a2aec83e04.zip
New upstream version 1.5.0+dfsgupstream/1.5.0+dfsg
Diffstat (limited to '')
-rw-r--r--python.d/varnish.chart.py239
1 files changed, 239 insertions, 0 deletions
diff --git a/python.d/varnish.chart.py b/python.d/varnish.chart.py
new file mode 100644
index 000000000..2b1512f4e
--- /dev/null
+++ b/python.d/varnish.chart.py
@@ -0,0 +1,239 @@
+# -*- coding: utf-8 -*-
+# Description: varnish netdata python.d module
+# Author: l2isbad
+
+from base import SimpleService
+from re import compile
+from os import access as is_executable, X_OK
+from subprocess import Popen, PIPE
+
+
+# default module values (can be overridden per job in `config`)
+# update_every = 2
+priority = 60000
+retries = 60
+
+ORDER = ['session', 'hit_rate', 'chit_rate', 'expunge', 'threads', 'backend_health', 'memory_usage', 'bad', 'uptime']
+
+CHARTS = {'backend_health':
+ {'lines': [['backend_conn', 'conn', 'incremental', 1, 1],
+ ['backend_unhealthy', 'unhealthy', 'incremental', 1, 1],
+ ['backend_busy', 'busy', 'incremental', 1, 1],
+ ['backend_fail', 'fail', 'incremental', 1, 1],
+ ['backend_reuse', 'reuse', 'incremental', 1, 1],
+ ['backend_recycle', 'resycle', 'incremental', 1, 1],
+ ['backend_toolate', 'toolate', 'incremental', 1, 1],
+ ['backend_retry', 'retry', 'incremental', 1, 1],
+ ['backend_req', 'req', 'incremental', 1, 1]],
+ 'options': [None, 'Backend health', 'connections', 'Backend health', 'varnish.backend_traf', 'line']},
+ 'bad':
+ {'lines': [['sess_drop_b', None, 'incremental', 1, 1],
+ ['backend_unhealthy_b', None, 'incremental', 1, 1],
+ ['fetch_failed', None, 'incremental', 1, 1],
+ ['backend_busy_b', None, 'incremental', 1, 1],
+ ['threads_failed_b', None, 'incremental', 1, 1],
+ ['threads_limited_b', None, 'incremental', 1, 1],
+ ['threads_destroyed_b', None, 'incremental', 1, 1],
+ ['thread_queue_len_b', 'queue_len', 'absolute', 1, 1],
+ ['losthdr_b', None, 'incremental', 1, 1],
+ ['esi_errors_b', None, 'incremental', 1, 1],
+ ['esi_warnings_b', None, 'incremental', 1, 1],
+ ['sess_fail_b', None, 'incremental', 1, 1],
+ ['sc_pipe_overflow_b', None, 'incremental', 1, 1],
+ ['sess_pipe_overflow_b', None, 'incremental', 1, 1]],
+ 'options': [None, 'Misbehavior', 'problems', 'Problems summary', 'varnish.bad', 'line']},
+ 'expunge':
+ {'lines': [['n_expired', 'expired', 'incremental', 1, 1],
+ ['n_lru_nuked', 'lru_nuked', 'incremental', 1, 1]],
+ 'options': [None, 'Object expunging', 'objects', 'Cache performance', 'varnish.expunge', 'line']},
+ 'hit_rate':
+ {'lines': [['cache_hit_perc', 'hit', 'absolute', 1, 100],
+ ['cache_miss_perc', 'miss', 'absolute', 1, 100],
+ ['cache_hitpass_perc', 'hitpass', 'absolute', 1, 100]],
+ 'options': [None, 'All history hit rate ratio','percent', 'Cache performance', 'varnish.hit_rate', 'stacked']},
+ 'chit_rate':
+ {'lines': [['cache_hit_cperc', 'hit', 'absolute', 1, 100],
+ ['cache_miss_cperc', 'miss', 'absolute', 1, 100],
+ ['cache_hitpass_cperc', 'hitpass', 'absolute', 1, 100]],
+ 'options': [None, 'Current poll hit rate ratio','percent', 'Cache performance', 'varnish.chit_rate', 'stacked']},
+ 'memory_usage':
+ {'lines': [['s0.g_space', 'available', 'absolute', 1, 1048576],
+ ['s0.g_bytes', 'allocated', 'absolute', -1, 1048576]],
+ 'options': [None, 'Memory usage', 'megabytes', 'Memory usage', 'varnish.memory_usage', 'stacked']},
+ 'session':
+ {'lines': [['sess_conn', 'sess_conn', 'incremental', 1, 1],
+ ['client_req', 'client_requests', 'incremental', 1, 1],
+ ['client_conn', 'client_conn', 'incremental', 1, 1],
+ ['client_drop', 'client_drop', 'incremental', 1, 1],
+ ['sess_dropped', 'sess_dropped', 'incremental', 1, 1]],
+ 'options': [None, 'Sessions', 'units', 'Client metrics', 'varnish.session', 'line']},
+ 'threads':
+ {'lines': [['threads', None, 'absolute', 1, 1],
+ ['threads_created', 'created', 'incremental', 1, 1],
+ ['threads_failed', 'failed', 'incremental', 1, 1],
+ ['threads_limited', 'limited', 'incremental', 1, 1],
+ ['thread_queue_len', 'queue_len', 'incremental', 1, 1],
+ ['sess_queued', 'sess_queued', 'incremental', 1, 1]],
+ 'options': [None, 'Thread status', 'threads', 'Thread-related metrics', 'varnish.threads', 'line']},
+ 'uptime':
+ {'lines': [['uptime', None, 'absolute', 1, 1]],
+ 'options': [None, 'Varnish uptime', 'seconds', 'Uptime', 'varnish.uptime', 'line']}
+}
+
+DIRECTORIES = ['/bin/', '/usr/bin/', '/sbin/', '/usr/sbin/']
+
+
+class Service(SimpleService):
+ def __init__(self, configuration=None, name=None):
+ SimpleService.__init__(self, configuration=configuration, name=name)
+ try:
+ self.varnish = [''.join([directory, 'varnishstat']) for directory in DIRECTORIES
+ if is_executable(''.join([directory, 'varnishstat']), X_OK)][0]
+ except IndexError:
+ self.varnish = False
+ self.rgx_all = compile(r'([A-Z]+\.)?([\d\w_.]+)\s+(\d+)')
+ # Could be
+ # VBE.boot.super_backend.pipe_hdrbyte (new)
+ # or
+ # VBE.default2(127.0.0.2,,81).bereq_bodybytes (old)
+ # Regex result: [('super_backend', 'beresp_hdrbytes', '0'), ('super_backend', 'beresp_bodybytes', '0')]
+ self.rgx_bck = (compile(r'VBE.([\d\w_.]+)\(.*?\).(beresp[\w_]+)\s+(\d+)'),
+ compile(r'VBE\.[\d\w-]+\.([\w\d_]+).(beresp[\w_]+)\s+(\d+)'))
+ self.cache_prev = list()
+
+ def check(self):
+ # Cant start without 'varnishstat' command
+ if not self.varnish:
+ self.error('\'varnishstat\' command was not found in %s or not executable by netdata' % DIRECTORIES)
+ return False
+
+ # If command is present and we can execute it we need to make sure..
+ # 1. STDOUT is not empty
+ reply = self._get_raw_data()
+ if not reply:
+ self.error('No output from \'varnishstat\' (not enough privileges?)')
+ return False
+
+ # 2. Output is parsable (list is not empty after regex findall)
+ is_parsable = self.rgx_all.findall(reply)
+ if not is_parsable:
+ self.error('Cant parse output...')
+ return False
+
+ # We need to find the right regex for backend parse
+ self.backend_list = self.rgx_bck[0].findall(reply)[::2]
+ if self.backend_list:
+ self.rgx_bck = self.rgx_bck[0]
+ else:
+ self.backend_list = self.rgx_bck[1].findall(reply)[::2]
+ self.rgx_bck = self.rgx_bck[1]
+
+ # We are about to start!
+ self.create_charts()
+
+ self.info('Plugin was started successfully')
+ return True
+
+ def _get_raw_data(self):
+ try:
+ reply = Popen([self.varnish, '-1'], stdout=PIPE, stderr=PIPE, shell=False)
+ except OSError:
+ return None
+
+ raw_data = reply.communicate()[0]
+
+ if not raw_data:
+ return None
+
+ return raw_data
+
+ def _get_data(self):
+ """
+ Format data received from shell command
+ :return: dict
+ """
+ raw_data = self._get_raw_data()
+ data_all = self.rgx_all.findall(raw_data)
+ data_backend = self.rgx_bck.findall(raw_data)
+
+ if not data_all:
+ return None
+
+ # 1. ALL data from 'varnishstat -1'. t - type(MAIN, MEMPOOL etc)
+ to_netdata = {k: int(v) for t, k, v in data_all}
+
+ # 2. ADD backend statistics
+ to_netdata.update({'_'.join([n, k]): int(v) for n, k, v in data_backend})
+
+ # 3. ADD additional keys to dict
+ # 3.1 Cache hit/miss/hitpass OVERALL in percent
+ cache_summary = sum([to_netdata.get('cache_hit', 0), to_netdata.get('cache_miss', 0),
+ to_netdata.get('cache_hitpass', 0)])
+ to_netdata['cache_hit_perc'] = find_percent(to_netdata.get('cache_hit', 0), cache_summary, 10000)
+ to_netdata['cache_miss_perc'] = find_percent(to_netdata.get('cache_miss', 0), cache_summary, 10000)
+ to_netdata['cache_hitpass_perc'] = find_percent(to_netdata.get('cache_hitpass', 0), cache_summary, 10000)
+
+ # 3.2 Cache hit/miss/hitpass CURRENT in percent
+ if self.cache_prev:
+ cache_summary = sum([to_netdata.get('cache_hit', 0), to_netdata.get('cache_miss', 0),
+ to_netdata.get('cache_hitpass', 0)]) - sum(self.cache_prev)
+ to_netdata['cache_hit_cperc'] = find_percent(to_netdata.get('cache_hit', 0) - self.cache_prev[0], cache_summary, 10000)
+ to_netdata['cache_miss_cperc'] = find_percent(to_netdata.get('cache_miss', 0) - self.cache_prev[1], cache_summary, 10000)
+ to_netdata['cache_hitpass_cperc'] = find_percent(to_netdata.get('cache_hitpass', 0) - self.cache_prev[2], cache_summary, 10000)
+ else:
+ to_netdata['cache_hit_cperc'] = 0
+ to_netdata['cache_miss_cperc'] = 0
+ to_netdata['cache_hitpass_cperc'] = 0
+
+ self.cache_prev = [to_netdata.get('cache_hit', 0), to_netdata.get('cache_miss', 0), to_netdata.get('cache_hitpass', 0)]
+
+ # 3.3 Problems summary chart
+ for elem in ['backend_busy', 'backend_unhealthy', 'esi_errors', 'esi_warnings', 'losthdr', 'sess_drop', 'sc_pipe_overflow',
+ 'sess_fail', 'sess_pipe_overflow', 'threads_destroyed', 'threads_failed', 'threads_limited', 'thread_queue_len']:
+ if to_netdata.get(elem) is not None:
+ to_netdata[''.join([elem, '_b'])] = to_netdata.get(elem)
+
+ # Ready steady go!
+ return to_netdata
+
+ def create_charts(self):
+ # If 'all_charts' is true...ALL charts are displayed. If no only default + 'extra_charts'
+ #if self.configuration.get('all_charts'):
+ # self.order = EXTRA_ORDER
+ #else:
+ # try:
+ # extra_charts = list(filter(lambda chart: chart in EXTRA_ORDER, self.extra_charts.split()))
+ # except (AttributeError, NameError, ValueError):
+ # self.error('Extra charts disabled.')
+ # extra_charts = []
+
+ self.order = ORDER[:]
+ #self.order.extend(extra_charts)
+
+ # Create static charts
+ #self.definitions = {chart: values for chart, values in CHARTS.items() if chart in self.order}
+ self.definitions = CHARTS
+
+ # Create dynamic backend charts
+ if self.backend_list:
+ for backend in self.backend_list:
+ self.order.insert(0, ''.join([backend[0], '_resp_stats']))
+ self.definitions.update({''.join([backend[0], '_resp_stats']): {
+ 'options': [None,
+ '%s response statistics' % backend[0].capitalize(),
+ "kilobit/s",
+ 'Backend response',
+ 'varnish.backend',
+ 'area'],
+ 'lines': [[''.join([backend[0], '_beresp_hdrbytes']),
+ 'header', 'incremental', 8, 1000],
+ [''.join([backend[0], '_beresp_bodybytes']),
+ 'body', 'incremental', -8, 1000]]}})
+
+
+def find_percent(value1, value2, multiply):
+ # If value2 is 0 return 0
+ if not value2:
+ return 0
+ else:
+ return round(float(value1) / float(value2) * multiply)