summaryrefslogtreecommitdiffstats
path: root/python.d/mongodb.chart.py
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2018-11-07 12:22:44 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2018-11-07 12:22:44 +0000
commit1e6c93250172946eeb38e94a92a1fd12c9d3011e (patch)
tree8ca5e16dfc7ad6b3bf2738ca0a48408a950f8f7e /python.d/mongodb.chart.py
parentUpdate watch file (diff)
downloadnetdata-1e6c93250172946eeb38e94a92a1fd12c9d3011e.tar.xz
netdata-1e6c93250172946eeb38e94a92a1fd12c9d3011e.zip
Merging upstream version 1.11.0+dfsg.
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'python.d/mongodb.chart.py')
-rw-r--r--python.d/mongodb.chart.py673
1 files changed, 0 insertions, 673 deletions
diff --git a/python.d/mongodb.chart.py b/python.d/mongodb.chart.py
deleted file mode 100644
index 909a419da..000000000
--- a/python.d/mongodb.chart.py
+++ /dev/null
@@ -1,673 +0,0 @@
-# -*- coding: utf-8 -*-
-# Description: mongodb netdata python.d module
-# Author: l2isbad
-
-from copy import deepcopy
-from datetime import datetime
-from sys import exc_info
-
-try:
- from pymongo import MongoClient, ASCENDING, DESCENDING
- from pymongo.errors import PyMongoError
- PYMONGO = True
-except ImportError:
- PYMONGO = False
-
-from bases.FrameworkServices.SimpleService import SimpleService
-
-# default module values (can be overridden per job in `config`)
-# update_every = 2
-priority = 60000
-retries = 60
-
-REPL_SET_STATES = [
- ('1', 'primary'),
- ('8', 'down'),
- ('2', 'secondary'),
- ('3', 'recovering'),
- ('5', 'startup2'),
- ('4', 'fatal'),
- ('7', 'arbiter'),
- ('6', 'unknown'),
- ('9', 'rollback'),
- ('10', 'removed'),
- ('0', 'startup')]
-
-
-def multiply_by_100(value):
- return value * 100
-
-
-DEFAULT_METRICS = [
- ('opcounters.delete', None, None),
- ('opcounters.update', None, None),
- ('opcounters.insert', None, None),
- ('opcounters.query', None, None),
- ('opcounters.getmore', None, None),
- ('globalLock.activeClients.readers', 'activeClients_readers', None),
- ('globalLock.activeClients.writers', 'activeClients_writers', None),
- ('connections.available', 'connections_available', None),
- ('connections.current', 'connections_current', None),
- ('mem.mapped', None, None),
- ('mem.resident', None, None),
- ('mem.virtual', None, None),
- ('globalLock.currentQueue.readers', 'currentQueue_readers', None),
- ('globalLock.currentQueue.writers', 'currentQueue_writers', None),
- ('asserts.msg', None, None),
- ('asserts.regular', None, None),
- ('asserts.user', None, None),
- ('asserts.warning', None, None),
- ('extra_info.page_faults', None, None),
- ('metrics.record.moves', None, None),
- ('backgroundFlushing.average_ms', None, multiply_by_100),
- ('backgroundFlushing.last_ms', None, multiply_by_100),
- ('backgroundFlushing.flushes', None, multiply_by_100),
- ('metrics.cursor.timedOut', None, None),
- ('metrics.cursor.open.total', 'cursor_total', None),
- ('metrics.cursor.open.noTimeout', None, None),
- ('cursors.timedOut', None, None),
- ('cursors.totalOpen', 'cursor_total', None)
-]
-
-DUR = [
- ('dur.commits', None, None),
- ('dur.journaledMB', None, multiply_by_100)
-]
-
-WIREDTIGER = [
- ('wiredTiger.concurrentTransactions.read.available', 'wiredTigerRead_available', None),
- ('wiredTiger.concurrentTransactions.read.out', 'wiredTigerRead_out', None),
- ('wiredTiger.concurrentTransactions.write.available', 'wiredTigerWrite_available', None),
- ('wiredTiger.concurrentTransactions.write.out', 'wiredTigerWrite_out', None),
- ('wiredTiger.cache.bytes currently in the cache', None, None),
- ('wiredTiger.cache.tracked dirty bytes in the cache', None, None),
- ('wiredTiger.cache.maximum bytes configured', None, None),
- ('wiredTiger.cache.unmodified pages evicted', 'unmodified', None),
- ('wiredTiger.cache.modified pages evicted', 'modified', None)
-]
-
-TCMALLOC = [
- ('tcmalloc.generic.current_allocated_bytes', None, None),
- ('tcmalloc.generic.heap_size', None, None),
- ('tcmalloc.tcmalloc.central_cache_free_bytes', None, None),
- ('tcmalloc.tcmalloc.current_total_thread_cache_bytes', None, None),
- ('tcmalloc.tcmalloc.pageheap_free_bytes', None, None),
- ('tcmalloc.tcmalloc.pageheap_unmapped_bytes', None, None),
- ('tcmalloc.tcmalloc.thread_cache_free_bytes', None, None),
- ('tcmalloc.tcmalloc.transfer_cache_free_bytes', None, None)
-]
-
-COMMANDS = [
- ('metrics.commands.count.total', 'count_total', None),
- ('metrics.commands.createIndexes.total', 'createIndexes_total', None),
- ('metrics.commands.delete.total', 'delete_total', None),
- ('metrics.commands.eval.total', 'eval_total', None),
- ('metrics.commands.findAndModify.total', 'findAndModify_total', None),
- ('metrics.commands.insert.total', 'insert_total', None),
- ('metrics.commands.delete.total', 'delete_total', None),
- ('metrics.commands.count.failed', 'count_failed', None),
- ('metrics.commands.createIndexes.failed', 'createIndexes_failed', None),
- ('metrics.commands.delete.failed', 'delete_failed', None),
- ('metrics.commands.eval.failed', 'eval_failed', None),
- ('metrics.commands.findAndModify.failed', 'findAndModify_failed', None),
- ('metrics.commands.insert.failed', 'insert_failed', None),
- ('metrics.commands.delete.failed', 'delete_failed', None)
-]
-
-LOCKS = [
- ('locks.Collection.acquireCount.R', 'Collection_R', None),
- ('locks.Collection.acquireCount.r', 'Collection_r', None),
- ('locks.Collection.acquireCount.W', 'Collection_W', None),
- ('locks.Collection.acquireCount.w', 'Collection_w', None),
- ('locks.Database.acquireCount.R', 'Database_R', None),
- ('locks.Database.acquireCount.r', 'Database_r', None),
- ('locks.Database.acquireCount.W', 'Database_W', None),
- ('locks.Database.acquireCount.w', 'Database_w', None),
- ('locks.Global.acquireCount.R', 'Global_R', None),
- ('locks.Global.acquireCount.r', 'Global_r', None),
- ('locks.Global.acquireCount.W', 'Global_W', None),
- ('locks.Global.acquireCount.w', 'Global_w', None),
- ('locks.Metadata.acquireCount.R', 'Metadata_R', None),
- ('locks.Metadata.acquireCount.w', 'Metadata_w', None),
- ('locks.oplog.acquireCount.r', 'oplog_r', None),
- ('locks.oplog.acquireCount.w', 'oplog_w', None)
-]
-
-DBSTATS = [
- 'dataSize',
- 'indexSize',
- 'storageSize',
- 'objects'
-]
-
-# charts order (can be overridden if you want less charts, or different order)
-ORDER = ['read_operations', 'write_operations', 'active_clients', 'journaling_transactions',
- 'journaling_volume', 'background_flush_average', 'background_flush_last', 'background_flush_rate',
- 'wiredtiger_read', 'wiredtiger_write', 'cursors', 'connections', 'memory', 'page_faults',
- 'queued_requests', 'record_moves', 'wiredtiger_cache', 'wiredtiger_pages_evicted', 'asserts',
- 'locks_collection', 'locks_database', 'locks_global', 'locks_metadata', 'locks_oplog',
- 'dbstats_objects', 'tcmalloc_generic', 'tcmalloc_metrics', 'command_total_rate', 'command_failed_rate']
-
-CHARTS = {
- 'read_operations': {
- 'options': [None, 'Received read requests', 'requests/s', 'throughput metrics',
- 'mongodb.read_operations', 'line'],
- 'lines': [
- ['query', None, 'incremental'],
- ['getmore', None, 'incremental']
- ]},
- 'write_operations': {
- 'options': [None, 'Received write requests', 'requests/s', 'throughput metrics',
- 'mongodb.write_operations', 'line'],
- 'lines': [
- ['insert', None, 'incremental'],
- ['update', None, 'incremental'],
- ['delete', None, 'incremental']
- ]},
- 'active_clients': {
- 'options': [None, 'Clients with read or write operations in progress or queued', 'clients',
- 'throughput metrics', 'mongodb.active_clients', 'line'],
- 'lines': [
- ['activeClients_readers', 'readers', 'absolute'],
- ['activeClients_writers', 'writers', 'absolute']
- ]},
- 'journaling_transactions': {
- 'options': [None, 'Transactions that have been written to the journal', 'commits',
- 'database performance', 'mongodb.journaling_transactions', 'line'],
- 'lines': [
- ['commits', None, 'absolute']
- ]},
- 'journaling_volume': {
- 'options': [None, 'Volume of data written to the journal', 'MB', 'database performance',
- 'mongodb.journaling_volume', 'line'],
- 'lines': [
- ['journaledMB', 'volume', 'absolute', 1, 100]
- ]},
- 'background_flush_average': {
- 'options': [None, 'Average time taken by flushes to execute', 'ms', 'database performance',
- 'mongodb.background_flush_average', 'line'],
- 'lines': [
- ['average_ms', 'time', 'absolute', 1, 100]
- ]},
- 'background_flush_last': {
- 'options': [None, 'Time taken by the last flush operation to execute', 'ms', 'database performance',
- 'mongodb.background_flush_last', 'line'],
- 'lines': [
- ['last_ms', 'time', 'absolute', 1, 100]
- ]},
- 'background_flush_rate': {
- 'options': [None, 'Flushes rate', 'flushes', 'database performance', 'mongodb.background_flush_rate', 'line'],
- 'lines': [
- ['flushes', 'flushes', 'incremental', 1, 1]
- ]},
- 'wiredtiger_read': {
- 'options': [None, 'Read tickets in use and remaining', 'tickets', 'database performance',
- 'mongodb.wiredtiger_read', 'stacked'],
- 'lines': [
- ['wiredTigerRead_available', 'available', 'absolute', 1, 1],
- ['wiredTigerRead_out', 'inuse', 'absolute', 1, 1]
- ]},
- 'wiredtiger_write': {
- 'options': [None, 'Write tickets in use and remaining', 'tickets', 'database performance',
- 'mongodb.wiredtiger_write', 'stacked'],
- 'lines': [
- ['wiredTigerWrite_available', 'available', 'absolute', 1, 1],
- ['wiredTigerWrite_out', 'inuse', 'absolute', 1, 1]
- ]},
- 'cursors': {
- 'options': [None, 'Currently openned cursors, cursors with timeout disabled and timed out cursors',
- 'cursors', 'database performance', 'mongodb.cursors', 'stacked'],
- 'lines': [
- ['cursor_total', 'openned', 'absolute', 1, 1],
- ['noTimeout', None, 'absolute', 1, 1],
- ['timedOut', None, 'incremental', 1, 1]
- ]},
- 'connections': {
- 'options': [None, 'Currently connected clients and unused connections', 'connections',
- 'resource utilization', 'mongodb.connections', 'stacked'],
- 'lines': [
- ['connections_available', 'unused', 'absolute', 1, 1],
- ['connections_current', 'connected', 'absolute', 1, 1]
- ]},
- 'memory': {
- 'options': [None, 'Memory metrics', 'MB', 'resource utilization', 'mongodb.memory', 'stacked'],
- 'lines': [
- ['virtual', None, 'absolute', 1, 1],
- ['resident', None, 'absolute', 1, 1],
- ['nonmapped', None, 'absolute', 1, 1],
- ['mapped', None, 'absolute', 1, 1]
- ]},
- 'page_faults': {
- 'options': [None, 'Number of times MongoDB had to fetch data from disk', 'request/s',
- 'resource utilization', 'mongodb.page_faults', 'line'],
- 'lines': [
- ['page_faults', None, 'incremental', 1, 1]
- ]},
- 'queued_requests': {
- 'options': [None, 'Currently queued read and wrire requests', 'requests', 'resource saturation',
- 'mongodb.queued_requests', 'line'],
- 'lines': [
- ['currentQueue_readers', 'readers', 'absolute', 1, 1],
- ['currentQueue_writers', 'writers', 'absolute', 1, 1]
- ]},
- 'record_moves': {
- 'options': [None, 'Number of times documents had to be moved on-disk', 'number',
- 'resource saturation', 'mongodb.record_moves', 'line'],
- 'lines': [
- ['moves', None, 'incremental', 1, 1]
- ]},
- 'asserts': {
- 'options': [None, 'Number of message, warning, regular, corresponding to errors generated'
- ' by users assertions raised', 'number', 'errors (asserts)', 'mongodb.asserts', 'line'],
- 'lines': [
- ['msg', None, 'incremental', 1, 1],
- ['warning', None, 'incremental', 1, 1],
- ['regular', None, 'incremental', 1, 1],
- ['user', None, 'incremental', 1, 1]
- ]},
- 'wiredtiger_cache': {
- 'options': [None, 'The percentage of the wiredTiger cache that is in use and cache with dirty bytes',
- 'percent', 'resource utilization', 'mongodb.wiredtiger_cache', 'stacked'],
- 'lines': [
- ['wiredTiger_percent_clean', 'inuse', 'absolute', 1, 1000],
- ['wiredTiger_percent_dirty', 'dirty', 'absolute', 1, 1000]
- ]},
- 'wiredtiger_pages_evicted': {
- 'options': [None, 'Pages evicted from the cache',
- 'pages', 'resource utilization', 'mongodb.wiredtiger_pages_evicted', 'stacked'],
- 'lines': [
- ['unmodified', None, 'absolute', 1, 1],
- ['modified', None, 'absolute', 1, 1]
- ]},
- 'dbstats_objects': {
- 'options': [None, 'Number of documents in the database among all the collections', 'documents',
- 'storage size metrics', 'mongodb.dbstats_objects', 'stacked'],
- 'lines': [
- ]},
- 'tcmalloc_generic': {
- 'options': [None, 'Tcmalloc generic metrics', 'MB', 'tcmalloc', 'mongodb.tcmalloc_generic', 'stacked'],
- 'lines': [
- ['current_allocated_bytes', 'allocated', 'absolute', 1, 1048576],
- ['heap_size', 'heap_size', 'absolute', 1, 1048576]
- ]},
- 'tcmalloc_metrics': {
- 'options': [None, 'Tcmalloc metrics', 'KB', 'tcmalloc', 'mongodb.tcmalloc_metrics', 'stacked'],
- 'lines': [
- ['central_cache_free_bytes', 'central_cache_free', 'absolute', 1, 1024],
- ['current_total_thread_cache_bytes', 'current_total_thread_cache', 'absolute', 1, 1024],
- ['pageheap_free_bytes', 'pageheap_free', 'absolute', 1, 1024],
- ['pageheap_unmapped_bytes', 'pageheap_unmapped', 'absolute', 1, 1024],
- ['thread_cache_free_bytes', 'thread_cache_free', 'absolute', 1, 1024],
- ['transfer_cache_free_bytes', 'transfer_cache_free', 'absolute', 1, 1024]
- ]},
- 'command_total_rate': {
- 'options': [None, 'Commands total rate', 'commands/s', 'commands', 'mongodb.command_total_rate', 'stacked'],
- 'lines': [
- ['count_total', 'count', 'incremental', 1, 1],
- ['createIndexes_total', 'createIndexes', 'incremental', 1, 1],
- ['delete_total', 'delete', 'incremental', 1, 1],
- ['eval_total', 'eval', 'incremental', 1, 1],
- ['findAndModify_total', 'findAndModify', 'incremental', 1, 1],
- ['insert_total', 'insert', 'incremental', 1, 1],
- ['update_total', 'update', 'incremental', 1, 1]
- ]},
- 'command_failed_rate': {
- 'options': [None, 'Commands failed rate', 'commands/s', 'commands', 'mongodb.command_failed_rate', 'stacked'],
- 'lines': [
- ['count_failed', 'count', 'incremental', 1, 1],
- ['createIndexes_failed', 'createIndexes', 'incremental', 1, 1],
- ['delete_failed', 'delete', 'incremental', 1, 1],
- ['eval_failed', 'eval', 'incremental', 1, 1],
- ['findAndModify_failed', 'findAndModify', 'incremental', 1, 1],
- ['insert_failed', 'insert', 'incremental', 1, 1],
- ['update_failed', 'update', 'incremental', 1, 1]
- ]},
- 'locks_collection': {
- 'options': [None, 'Collection lock. Number of times the lock was acquired in the specified mode',
- 'locks', 'locks metrics', 'mongodb.locks_collection', 'stacked'],
- 'lines': [
- ['Collection_R', 'shared', 'incremental'],
- ['Collection_W', 'exclusive', 'incremental'],
- ['Collection_r', 'intent_shared', 'incremental'],
- ['Collection_w', 'intent_exclusive', 'incremental']
- ]},
- 'locks_database': {
- 'options': [None, 'Database lock. Number of times the lock was acquired in the specified mode',
- 'locks', 'locks metrics', 'mongodb.locks_database', 'stacked'],
- 'lines': [
- ['Database_R', 'shared', 'incremental'],
- ['Database_W', 'exclusive', 'incremental'],
- ['Database_r', 'intent_shared', 'incremental'],
- ['Database_w', 'intent_exclusive', 'incremental']
- ]},
- 'locks_global': {
- 'options': [None, 'Global lock. Number of times the lock was acquired in the specified mode',
- 'locks', 'locks metrics', 'mongodb.locks_global', 'stacked'],
- 'lines': [
- ['Global_R', 'shared', 'incremental'],
- ['Global_W', 'exclusive', 'incremental'],
- ['Global_r', 'intent_shared', 'incremental'],
- ['Global_w', 'intent_exclusive', 'incremental']
- ]},
- 'locks_metadata': {
- 'options': [None, 'Metadata lock. Number of times the lock was acquired in the specified mode',
- 'locks', 'locks metrics', 'mongodb.locks_metadata', 'stacked'],
- 'lines': [
- ['Metadata_R', 'shared', 'incremental'],
- ['Metadata_w', 'intent_exclusive', 'incremental']
- ]},
- 'locks_oplog': {
- 'options': [None, 'Lock on the oplog. Number of times the lock was acquired in the specified mode',
- 'locks', 'locks metrics', 'mongodb.locks_oplog', 'stacked'],
- 'lines': [
- ['oplog_r', 'intent_shared', 'incremental'],
- ['oplog_w', 'intent_exclusive', 'incremental']
- ]}
-}
-
-
-class Service(SimpleService):
- def __init__(self, configuration=None, name=None):
- SimpleService.__init__(self, configuration=configuration, name=name)
- self.order = ORDER[:]
- self.definitions = deepcopy(CHARTS)
- self.user = self.configuration.get('user')
- self.password = self.configuration.get('pass')
- self.host = self.configuration.get('host', '127.0.0.1')
- self.port = self.configuration.get('port', 27017)
- self.timeout = self.configuration.get('timeout', 100)
- self.metrics_to_collect = deepcopy(DEFAULT_METRICS)
- self.connection = None
- self.do_replica = None
- self.databases = list()
-
- def check(self):
- if not PYMONGO:
- self.error('Pymongo module is needed to use mongodb.chart.py')
- return False
- self.connection, server_status, error = self._create_connection()
- if error:
- self.error(error)
- return False
-
- self.build_metrics_to_collect_(server_status)
-
- try:
- data = self._get_data()
- except (LookupError, SyntaxError, AttributeError):
- self.error('Type: %s, error: %s' % (str(exc_info()[0]), str(exc_info()[1])))
- return False
- if isinstance(data, dict) and data:
- self._data_from_check = data
- self.create_charts_(server_status)
- return True
- self.error('_get_data() returned no data or type is not <dict>')
- return False
-
- def build_metrics_to_collect_(self, server_status):
-
- self.do_replica = 'repl' in server_status
- if 'dur' in server_status:
- self.metrics_to_collect.extend(DUR)
- if 'tcmalloc' in server_status:
- self.metrics_to_collect.extend(TCMALLOC)
- if 'commands' in server_status['metrics']:
- self.metrics_to_collect.extend(COMMANDS)
- if 'wiredTiger' in server_status:
- self.metrics_to_collect.extend(WIREDTIGER)
- if 'Collection' in server_status['locks']:
- self.metrics_to_collect.extend(LOCKS)
-
- def create_charts_(self, server_status):
-
- if 'dur' not in server_status:
- self.order.remove('journaling_transactions')
- self.order.remove('journaling_volume')
-
- if 'backgroundFlushing' not in server_status:
- self.order.remove('background_flush_average')
- self.order.remove('background_flush_last')
- self.order.remove('background_flush_rate')
-
- if 'wiredTiger' not in server_status:
- self.order.remove('wiredtiger_write')
- self.order.remove('wiredtiger_read')
- self.order.remove('wiredtiger_cache')
-
- if 'tcmalloc' not in server_status:
- self.order.remove('tcmalloc_generic')
- self.order.remove('tcmalloc_metrics')
-
- if 'commands' not in server_status['metrics']:
- self.order.remove('command_total_rate')
- self.order.remove('command_failed_rate')
-
- if 'Collection' not in server_status['locks']:
- self.order.remove('locks_collection')
- self.order.remove('locks_database')
- self.order.remove('locks_global')
- self.order.remove('locks_metadata')
-
- if 'oplog' not in server_status['locks']:
- self.order.remove('locks_oplog')
-
- for dbase in self.databases:
- self.order.append('_'.join([dbase, 'dbstats']))
- self.definitions['_'.join([dbase, 'dbstats'])] = {
- 'options': [None, '%s: size of all documents, indexes, extents' % dbase, 'KB',
- 'storage size metrics', 'mongodb.dbstats', 'line'],
- 'lines': [
- ['_'.join([dbase, 'dataSize']), 'documents', 'absolute', 1, 1024],
- ['_'.join([dbase, 'indexSize']), 'indexes', 'absolute', 1, 1024],
- ['_'.join([dbase, 'storageSize']), 'extents', 'absolute', 1, 1024]
- ]}
- self.definitions['dbstats_objects']['lines'].append(['_'.join([dbase, 'objects']), dbase, 'absolute'])
-
- if self.do_replica:
- def create_lines(hosts, string):
- lines = list()
- for host in hosts:
- dim_id = '_'.join([host, string])
- lines.append([dim_id, host, 'absolute', 1, 1000])
- return lines
-
- def create_state_lines(states):
- lines = list()
- for state, description in states:
- dim_id = '_'.join([host, 'state', state])
- lines.append([dim_id, description, 'absolute', 1, 1])
- return lines
-
- all_hosts = server_status['repl']['hosts'] + server_status['repl'].get('arbiters', list())
- this_host = server_status['repl']['me']
- other_hosts = [host for host in all_hosts if host != this_host]
-
- if 'local' in self.databases:
- self.order.append('oplog_window')
- self.definitions['oplog_window'] = {
- 'options': [None, 'Interval of time between the oldest and the latest entries in the oplog',
- 'seconds', 'replication and oplog', 'mongodb.oplog_window', 'line'],
- 'lines': [['timeDiff', 'window', 'absolute', 1, 1000]]}
- # Create "heartbeat delay" chart
- self.order.append('heartbeat_delay')
- self.definitions['heartbeat_delay'] = {
- 'options': [None, 'Time when last heartbeat was received'
- ' from the replica set member (lastHeartbeatRecv)',
- 'seconds ago', 'replication and oplog', 'mongodb.replication_heartbeat_delay', 'stacked'],
- 'lines': create_lines(other_hosts, 'heartbeat_lag')}
- # Create "optimedate delay" chart
- self.order.append('optimedate_delay')
- self.definitions['optimedate_delay'] = {
- 'options': [None, 'Time when last entry from the oplog was applied (optimeDate)',
- 'seconds ago', 'replication and oplog', 'mongodb.replication_optimedate_delay', 'stacked'],
- 'lines': create_lines(all_hosts, 'optimedate')}
- # Create "replica set members state" chart
- for host in all_hosts:
- chart_name = '_'.join([host, 'state'])
- self.order.append(chart_name)
- self.definitions[chart_name] = {
- 'options': [None, 'Replica set member (%s) current state' % host, 'state',
- 'replication and oplog', 'mongodb.replication_state', 'line'],
- 'lines': create_state_lines(REPL_SET_STATES)}
-
- def _get_raw_data(self):
- raw_data = dict()
-
- raw_data.update(self.get_server_status() or dict())
- raw_data.update(self.get_db_stats() or dict())
- raw_data.update(self.get_repl_set_get_status() or dict())
- raw_data.update(self.get_get_replication_info() or dict())
-
- return raw_data or None
-
- def get_server_status(self):
- raw_data = dict()
- try:
- raw_data['serverStatus'] = self.connection.admin.command('serverStatus')
- except PyMongoError:
- return None
- else:
- return raw_data
-
- def get_db_stats(self):
- if not self.databases:
- return None
-
- raw_data = dict()
- raw_data['dbStats'] = dict()
- try:
- for dbase in self.databases:
- raw_data['dbStats'][dbase] = self.connection[dbase].command('dbStats')
- return raw_data
- except PyMongoError:
- return None
-
- def get_repl_set_get_status(self):
- if not self.do_replica:
- return None
-
- raw_data = dict()
- try:
- raw_data['replSetGetStatus'] = self.connection.admin.command('replSetGetStatus')
- return raw_data
- except PyMongoError:
- return None
-
- def get_get_replication_info(self):
- if not (self.do_replica and 'local' in self.databases):
- return None
-
- raw_data = dict()
- raw_data['getReplicationInfo'] = dict()
- try:
- raw_data['getReplicationInfo']['ASCENDING'] = self.connection.local.oplog.rs.find().sort(
- "$natural", ASCENDING).limit(1)[0]
- raw_data['getReplicationInfo']['DESCENDING'] = self.connection.local.oplog.rs.find().sort(
- "$natural", DESCENDING).limit(1)[0]
- return raw_data
- except PyMongoError:
- return None
-
- def _get_data(self):
- """
- :return: dict
- """
- raw_data = self._get_raw_data()
-
- if not raw_data:
- return None
-
- to_netdata = dict()
- serverStatus = raw_data['serverStatus']
- dbStats = raw_data.get('dbStats')
- replSetGetStatus = raw_data.get('replSetGetStatus')
- getReplicationInfo = raw_data.get('getReplicationInfo')
- utc_now = datetime.utcnow()
-
- # serverStatus
- for metric, new_name, func in self.metrics_to_collect:
- value = serverStatus
- for key in metric.split('.'):
- try:
- value = value[key]
- except KeyError:
- break
-
- if not isinstance(value, dict) and key:
- to_netdata[new_name or key] = value if not func else func(value)
-
- to_netdata['nonmapped'] = to_netdata['virtual'] - serverStatus['mem'].get('mappedWithJournal',
- to_netdata['mapped'])
- if to_netdata.get('maximum bytes configured'):
- maximum = to_netdata['maximum bytes configured']
- to_netdata['wiredTiger_percent_clean'] = int(to_netdata['bytes currently in the cache']
- * 100 / maximum * 1000)
- to_netdata['wiredTiger_percent_dirty'] = int(to_netdata['tracked dirty bytes in the cache']
- * 100 / maximum * 1000)
-
- # dbStats
- if dbStats:
- for dbase in dbStats:
- for metric in DBSTATS:
- key = '_'.join([dbase, metric])
- to_netdata[key] = dbStats[dbase][metric]
-
- # replSetGetStatus
- if replSetGetStatus:
- other_hosts = list()
- members = replSetGetStatus['members']
- unix_epoch = datetime(1970, 1, 1, 0, 0)
-
- for member in members:
- if not member.get('self'):
- other_hosts.append(member)
- # Replica set time diff between current time and time when last entry from the oplog was applied
- if member.get('optimeDate', unix_epoch) != unix_epoch:
- member_optimedate = member['name'] + '_optimedate'
- to_netdata.update({member_optimedate: int(delta_calculation(delta=utc_now - member['optimeDate'],
- multiplier=1000))})
- # Replica set members state
- member_state = member['name'] + '_state'
- for elem in REPL_SET_STATES:
- state = elem[0]
- to_netdata.update({'_'.join([member_state, state]): 0})
- to_netdata.update({'_'.join([member_state, str(member['state'])]): member['state']})
- # Heartbeat lag calculation
- for other in other_hosts:
- if other['lastHeartbeatRecv'] != unix_epoch:
- node = other['name'] + '_heartbeat_lag'
- to_netdata[node] = int(delta_calculation(delta=utc_now - other['lastHeartbeatRecv'],
- multiplier=1000))
-
- if getReplicationInfo:
- first_event = getReplicationInfo['ASCENDING']['ts'].as_datetime()
- last_event = getReplicationInfo['DESCENDING']['ts'].as_datetime()
- to_netdata['timeDiff'] = int(delta_calculation(delta=last_event - first_event, multiplier=1000))
-
- return to_netdata
-
- def _create_connection(self):
- conn_vars = {'host': self.host, 'port': self.port}
- if hasattr(MongoClient, 'server_selection_timeout'):
- conn_vars.update({'serverselectiontimeoutms': self.timeout})
- try:
- connection = MongoClient(**conn_vars)
- if self.user and self.password:
- connection.admin.authenticate(name=self.user, password=self.password)
- # elif self.user:
- # connection.admin.authenticate(name=self.user, mechanism='MONGODB-X509')
- server_status = connection.admin.command('serverStatus')
- except PyMongoError as error:
- return None, None, str(error)
- else:
- try:
- self.databases = connection.database_names()
- except PyMongoError as error:
- self.info('Can\'t collect databases: %s' % str(error))
- return connection, server_status, None
-
-
-def delta_calculation(delta, multiplier=1):
- if hasattr(delta, 'total_seconds'):
- return delta.total_seconds() * multiplier
- return (delta.microseconds + (delta.seconds + delta.days * 24 * 3600) * 10 ** 6) / 10.0 ** 6 * multiplier