summaryrefslogtreecommitdiffstats
path: root/python.d/elasticsearch.chart.py
diff options
context:
space:
mode:
authorLennart Weller <lhw@ring0.de>2017-09-17 22:17:33 +0000
committerLennart Weller <lhw@ring0.de>2017-09-17 22:17:33 +0000
commit6aaf5ba7ed0980c14bdc554fc8839a2126455ed5 (patch)
tree6161925716661486e7f47c479668a9487b039d83 /python.d/elasticsearch.chart.py
parentNew upstream version 1.7.0+dfsg (diff)
downloadnetdata-6aaf5ba7ed0980c14bdc554fc8839a2126455ed5.tar.xz
netdata-6aaf5ba7ed0980c14bdc554fc8839a2126455ed5.zip
New upstream version 1.8.0+dfsgupstream/1.8.0+dfsg
Diffstat (limited to 'python.d/elasticsearch.chart.py')
-rw-r--r--python.d/elasticsearch.chart.py417
1 files changed, 204 insertions, 213 deletions
diff --git a/python.d/elasticsearch.chart.py b/python.d/elasticsearch.chart.py
index 9ec08719..2e0f18c0 100644
--- a/python.d/elasticsearch.chart.py
+++ b/python.d/elasticsearch.chart.py
@@ -2,15 +2,16 @@
# Description: elastic search node stats netdata python.d module
# Author: l2isbad
-from base import UrlService
+from collections import namedtuple
+from json import loads
from socket import gethostbyname, gaierror
+from threading import Thread
try:
from queue import Queue
except ImportError:
from Queue import Queue
-from threading import Thread
-from collections import namedtuple
-from json import loads
+
+from base import UrlService
# default module values (can be overridden per job in `config`)
# update_every = 2
@@ -18,214 +19,216 @@ update_every = 5
priority = 60000
retries = 60
-METHODS = namedtuple('METHODS', ['get_data_function', 'url'])
+METHODS = namedtuple('METHODS', ['get_data', 'url', 'run'])
NODE_STATS = [
- ('indices.search.fetch_current', None, None),
- ('indices.search.fetch_total', None, None),
- ('indices.search.query_current', None, None),
- ('indices.search.query_total', None, None),
- ('indices.search.query_time_in_millis', None, None),
- ('indices.search.fetch_time_in_millis', None, None),
- ('indices.indexing.index_total', 'indexing_index_total', None),
- ('indices.indexing.index_current', 'indexing_index_current', None),
- ('indices.indexing.index_time_in_millis', 'indexing_index_time_in_millis', None),
- ('indices.refresh.total', 'refresh_total', None),
- ('indices.refresh.total_time_in_millis', 'refresh_total_time_in_millis', None),
- ('indices.flush.total', 'flush_total', None),
- ('indices.flush.total_time_in_millis', 'flush_total_time_in_millis', None),
- ('jvm.gc.collectors.young.collection_count', 'young_collection_count', None),
- ('jvm.gc.collectors.old.collection_count', 'old_collection_count', None),
- ('jvm.gc.collectors.young.collection_time_in_millis', 'young_collection_time_in_millis', None),
- ('jvm.gc.collectors.old.collection_time_in_millis', 'old_collection_time_in_millis', None),
- ('jvm.mem.heap_used_percent', 'jvm_heap_percent', None),
- ('jvm.mem.heap_committed_in_bytes', 'jvm_heap_commit', None),
- ('thread_pool.bulk.queue', 'bulk_queue', None),
- ('thread_pool.bulk.rejected', 'bulk_rejected', None),
- ('thread_pool.index.queue', 'index_queue', None),
- ('thread_pool.index.rejected', 'index_rejected', None),
- ('thread_pool.search.queue', 'search_queue', None),
- ('thread_pool.search.rejected', 'search_rejected', None),
- ('thread_pool.merge.queue', 'merge_queue', None),
- ('thread_pool.merge.rejected', 'merge_rejected', None),
- ('indices.fielddata.memory_size_in_bytes', 'index_fdata_memory', None),
- ('indices.fielddata.evictions', None, None),
- ('breakers.fielddata.tripped', None, None),
- ('http.current_open', 'http_current_open', None),
- ('transport.rx_size_in_bytes', 'transport_rx_size_in_bytes', None),
- ('transport.tx_size_in_bytes', 'transport_tx_size_in_bytes', None),
- ('process.max_file_descriptors', None, None),
- ('process.open_file_descriptors', None, None)
+ 'indices.search.fetch_current',
+ 'indices.search.fetch_total',
+ 'indices.search.query_current',
+ 'indices.search.query_total',
+ 'indices.search.query_time_in_millis',
+ 'indices.search.fetch_time_in_millis',
+ 'indices.indexing.index_total',
+ 'indices.indexing.index_current',
+ 'indices.indexing.index_time_in_millis',
+ 'indices.refresh.total',
+ 'indices.refresh.total_time_in_millis',
+ 'indices.flush.total'
+ 'indices.flush.total_time_in_millis',
+ 'jvm.gc.collectors.young.collection_count',
+ 'jvm.gc.collectors.old.collection_count',
+ 'jvm.gc.collectors.young.collection_time_in_millis',
+ 'jvm.gc.collectors.old.collection_time_in_millis',
+ 'jvm.mem.heap_used_percent',
+ 'jvm.mem.heap_committed_in_bytes',
+ 'thread_pool.bulk.queue'
+ 'thread_pool.bulk.rejected',
+ 'thread_pool.index.queue',
+ 'thread_pool.index.rejected',
+ 'thread_pool.search.queue',
+ 'thread_pool.search.rejected',
+ 'thread_pool.merge.queue',
+ 'thread_pool.merge.rejected',
+ 'indices.fielddata.memory_size_in_bytes',
+ 'indices.fielddata.evictions',
+ 'breakers.fielddata.tripped',
+ 'http.current_open',
+ 'transport.rx_size_in_bytes',
+ 'transport.tx_size_in_bytes',
+ 'process.max_file_descriptors',
+ 'process.open_file_descriptors'
]
CLUSTER_STATS = [
- ('nodes.count.data_only', 'count_data_only', None),
- ('nodes.count.master_data', 'count_master_data', None),
- ('nodes.count.total', 'count_total', None),
- ('nodes.count.master_only', 'count_master_only', None),
- ('nodes.count.client', 'count_client', None),
- ('indices.docs.count', 'docs_count', None),
- ('indices.query_cache.hit_count', 'query_cache_hit_count', None),
- ('indices.query_cache.miss_count', 'query_cache_miss_count', None),
- ('indices.store.size_in_bytes', 'store_size_in_bytes', None),
- ('indices.count', 'indices_count', None),
- ('indices.shards.total', 'shards_total', None)
+ 'nodes.count.data_only',
+ 'nodes.count.master_data',
+ 'nodes.count.total',
+ 'nodes.count.master_only',
+ 'nodes.count.client',
+ 'indices.docs.count',
+ 'indices.query_cache.hit_count',
+ 'indices.query_cache.miss_count',
+ 'indices.store.size_in_bytes',
+ 'indices.count',
+ 'indices.shards.total'
]
HEALTH_STATS = [
- ('number_of_nodes', 'health_number_of_nodes', None),
- ('number_of_data_nodes', 'health_number_of_data_nodes', None),
- ('number_of_pending_tasks', 'health_number_of_pending_tasks', None),
- ('number_of_in_flight_fetch', 'health_number_of_in_flight_fetch', None),
- ('active_shards', 'health_active_shards', None),
- ('relocating_shards', 'health_relocating_shards', None),
- ('unassigned_shards', 'health_unassigned_shards', None),
- ('delayed_unassigned_shards', 'health_delayed_unassigned_shards', None),
- ('initializing_shards', 'health_initializing_shards', None),
- ('active_shards_percent_as_number', 'health_active_shards_percent_as_number', None)
+ 'number_of_nodes',
+ 'number_of_data_nodes',
+ 'number_of_pending_tasks',
+ 'number_of_in_flight_fetch',
+ 'active_shards',
+ 'relocating_shards',
+ 'unassigned_shards',
+ 'delayed_unassigned_shards',
+ 'initializing_shards',
+ 'active_shards_percent_as_number'
]
LATENCY = {
'query_latency':
- {'total': 'query_total',
- 'spent_time': 'query_time_in_millis'},
+ {'total': 'indices_search_query_total',
+ 'spent_time': 'indices_search_query_time_in_millis'},
'fetch_latency':
- {'total': 'fetch_total',
- 'spent_time': 'fetch_time_in_millis'},
+ {'total': 'indices_search_fetch_total',
+ 'spent_time': 'indices_search_fetch_time_in_millis'},
'indexing_latency':
- {'total': 'indexing_index_total',
- 'spent_time': 'indexing_index_time_in_millis'},
+ {'total': 'indices_indexing_index_total',
+ 'spent_time': 'indices_indexing_index_time_in_millis'},
'flushing_latency':
- {'total': 'flush_total',
- 'spent_time': 'flush_total_time_in_millis'}
+ {'total': 'indices_flush_total',
+ 'spent_time': 'indices_flush_total_time_in_millis'}
}
# charts order (can be overridden if you want less charts, or different order)
-ORDER = ['search_perf_total', 'search_perf_current', 'search_perf_time', 'search_latency', 'index_perf_total',
- 'index_perf_current', 'index_perf_time', 'index_latency', 'jvm_mem_heap', 'jvm_gc_count',
- 'jvm_gc_time', 'host_metrics_file_descriptors', 'host_metrics_http', 'host_metrics_transport',
- 'thread_pool_qr_q', 'thread_pool_qr_r', 'fdata_cache', 'fdata_ev_tr', 'cluster_health_status',
- 'cluster_health_nodes', 'cluster_health_shards', 'cluster_stats_nodes', 'cluster_stats_query_cache',
- 'cluster_stats_docs', 'cluster_stats_store', 'cluster_stats_indices_shards']
+ORDER = ['search_performance_total', 'search_performance_current', 'search_performance_time',
+ 'search_latency', 'index_performance_total', 'index_performance_current', 'index_performance_time',
+ 'index_latency', 'jvm_mem_heap', 'jvm_gc_count', 'jvm_gc_time', 'host_metrics_file_descriptors',
+ 'host_metrics_http', 'host_metrics_transport', 'thread_pool_queued', 'thread_pool_rejected',
+ 'fielddata_cache', 'fielddata_evictions_tripped', 'cluster_health_status', 'cluster_health_nodes',
+ 'cluster_health_shards', 'cluster_stats_nodes', 'cluster_stats_query_cache', 'cluster_stats_docs',
+ 'cluster_stats_store', 'cluster_stats_indices_shards']
CHARTS = {
- 'search_perf_total': {
+ 'search_performance_total': {
'options': [None, 'Queries And Fetches', 'number of', 'search performance',
- 'es.search_query_total', 'stacked'],
+ 'elastic.search_performance_total', 'stacked'],
'lines': [
- ['query_total', 'queries', 'incremental'],
- ['fetch_total', 'fetches', 'incremental']
+ ['indices_search_query_total', 'queries', 'incremental'],
+ ['indices_search_fetch_total', 'fetches', 'incremental']
]},
- 'search_perf_current': {
+ 'search_performance_current': {
'options': [None, 'Queries and Fetches In Progress', 'number of', 'search performance',
- 'es.search_query_current', 'stacked'],
+ 'elastic.search_performance_current', 'stacked'],
'lines': [
- ['query_current', 'queries', 'absolute'],
- ['fetch_current', 'fetches', 'absolute']
+ ['indices_search_query_current', 'queries', 'absolute'],
+ ['indices_search_fetch_current', 'fetches', 'absolute']
]},
- 'search_perf_time': {
+ 'search_performance_time': {
'options': [None, 'Time Spent On Queries And Fetches', 'seconds', 'search performance',
- 'es.search_time', 'stacked'],
+ 'elastic.search_performance_time', 'stacked'],
'lines': [
- ['query_time_in_millis', 'query', 'incremental', 1, 1000],
- ['fetch_time_in_millis', 'fetch', 'incremental', 1, 1000]
+ ['indices_search_query_time_in_millis', 'query', 'incremental', 1, 1000],
+ ['indices_search_fetch_time_in_millis', 'fetch', 'incremental', 1, 1000]
]},
'search_latency': {
- 'options': [None, 'Query And Fetch Latency', 'ms', 'search performance', 'es.search_latency', 'stacked'],
+ 'options': [None, 'Query And Fetch Latency', 'ms', 'search performance', 'elastic.search_latency', 'stacked'],
'lines': [
['query_latency', 'query', 'absolute', 1, 1000],
['fetch_latency', 'fetch', 'absolute', 1, 1000]
]},
- 'index_perf_total': {
+ 'index_performance_total': {
'options': [None, 'Indexed Documents, Index Refreshes, Index Flushes To Disk', 'number of',
- 'indexing performance', 'es.index_performance_total', 'stacked'],
+ 'indexing performance', 'elastic.index_performance_total', 'stacked'],
'lines': [
- ['indexing_index_total', 'indexed', 'incremental'],
- ['refresh_total', 'refreshes', 'incremental'],
- ['flush_total', 'flushes', 'incremental']
+ ['indices_indexing_index_total', 'indexed', 'incremental'],
+ ['indices_refresh_total', 'refreshes', 'incremental'],
+ ['indices_flush_total', 'flushes', 'incremental']
]},
- 'index_perf_current': {
+ 'index_performance_current': {
'options': [None, 'Number Of Documents Currently Being Indexed', 'currently indexed',
- 'indexing performance', 'es.index_performance_current', 'stacked'],
+ 'indexing performance', 'elastic.index_performance_current', 'stacked'],
'lines': [
- ['indexing_index_current', 'documents', 'absolute']
+ ['indices_indexing_index_current', 'documents', 'absolute']
]},
- 'index_perf_time': {
+ 'index_performance_time': {
'options': [None, 'Time Spent On Indexing, Refreshing, Flushing', 'seconds', 'indexing performance',
- 'es.search_time', 'stacked'],
+ 'elastic.index_performance_time', 'stacked'],
'lines': [
- ['indexing_index_time_in_millis', 'indexing', 'incremental', 1, 1000],
- ['refresh_total_time_in_millis', 'refreshing', 'incremental', 1, 1000],
- ['flush_total_time_in_millis', 'flushing', 'incremental', 1, 1000]
+ ['indices_indexing_index_time_in_millis', 'indexing', 'incremental', 1, 1000],
+ ['indices_refresh_total_time_in_millis', 'refreshing', 'incremental', 1, 1000],
+ ['indices_flush_total_time_in_millis', 'flushing', 'incremental', 1, 1000]
]},
'index_latency': {
'options': [None, 'Indexing And Flushing Latency', 'ms', 'indexing performance',
- 'es.index_latency', 'stacked'],
+ 'elastic.index_latency', 'stacked'],
'lines': [
['indexing_latency', 'indexing', 'absolute', 1, 1000],
['flushing_latency', 'flushing', 'absolute', 1, 1000]
]},
'jvm_mem_heap': {
'options': [None, 'JVM Heap Currently in Use/Committed', 'percent/MB', 'memory usage and gc',
- 'es.jvm_heap', 'area'],
+ 'elastic.jvm_heap', 'area'],
'lines': [
- ['jvm_heap_percent', 'inuse', 'absolute'],
- ['jvm_heap_commit', 'commit', 'absolute', -1, 1048576]
+ ['jvm_mem_heap_used_percent', 'inuse', 'absolute'],
+ ['jvm_mem_heap_committed_in_bytes', 'commit', 'absolute', -1, 1048576]
]},
'jvm_gc_count': {
- 'options': [None, 'Garbage Collections', 'counts', 'memory usage and gc', 'es.gc_count', 'stacked'],
+ 'options': [None, 'Garbage Collections', 'counts', 'memory usage and gc', 'elastic.gc_count', 'stacked'],
'lines': [
- ['young_collection_count', 'young', 'incremental'],
- ['old_collection_count', 'old', 'incremental']
+ ['jvm_gc_collectors_young_collection_count', 'young', 'incremental'],
+ ['jvm_gc_collectors_old_collection_count', 'old', 'incremental']
]},
'jvm_gc_time': {
- 'options': [None, 'Time Spent On Garbage Collections', 'ms', 'memory usage and gc', 'es.gc_time', 'stacked'],
+ 'options': [None, 'Time Spent On Garbage Collections', 'ms', 'memory usage and gc',
+ 'elastic.gc_time', 'stacked'],
'lines': [
- ['young_collection_time_in_millis', 'young', 'incremental'],
- ['old_collection_time_in_millis', 'old', 'incremental']
+ ['jvm_gc_collectors_young_collection_time_in_millis', 'young', 'incremental'],
+ ['jvm_gc_collectors_old_collection_time_in_millis', 'old', 'incremental']
]},
- 'thread_pool_qr_q': {
+ 'thread_pool_queued': {
'options': [None, 'Number Of Queued Threads In Thread Pool', 'queued threads', 'queues and rejections',
- 'es.thread_pool_queued', 'stacked'],
+ 'elastic.thread_pool_queued', 'stacked'],
'lines': [
- ['bulk_queue', 'bulk', 'absolute'],
- ['index_queue', 'index', 'absolute'],
- ['search_queue', 'search', 'absolute'],
- ['merge_queue', 'merge', 'absolute']
+ ['thread_pool_bulk_queue', 'bulk', 'absolute'],
+ ['thread_pool_index_queue', 'index', 'absolute'],
+ ['thread_pool_search_queue', 'search', 'absolute'],
+ ['thread_pool_merge_queue', 'merge', 'absolute']
]},
- 'thread_pool_qr_r': {
+ 'thread_pool_rejected': {
'options': [None, 'Rejected Threads In Thread Pool', 'rejected threads', 'queues and rejections',
- 'es.thread_pool_rejected', 'stacked'],
+ 'elastic.thread_pool_rejected', 'stacked'],
'lines': [
- ['bulk_rejected', 'bulk', 'absolute'],
- ['index_rejected', 'index', 'absolute'],
- ['search_rejected', 'search', 'absolute'],
- ['merge_rejected', 'merge', 'absolute']
+ ['thread_pool_bulk_rejected', 'bulk', 'absolute'],
+ ['thread_pool_index_rejected', 'index', 'absolute'],
+ ['thread_pool_search_rejected', 'search', 'absolute'],
+ ['thread_pool_merge_rejected', 'merge', 'absolute']
]},
- 'fdata_cache': {
- 'options': [None, 'Fielddata Cache', 'MB', 'fielddata cache', 'es.fdata_cache', 'line'],
+ 'fielddata_cache': {
+ 'options': [None, 'Fielddata Cache', 'MB', 'fielddata cache', 'elastic.fielddata_cache', 'line'],
'lines': [
- ['index_fdata_memory', 'cache', 'absolute', 1, 1048576]
+ ['indices_fielddata_memory_size_in_bytes', 'cache', 'absolute', 1, 1048576]
]},
- 'fdata_ev_tr': {
+ 'fielddata_evictions_tripped': {
'options': [None, 'Fielddata Evictions And Circuit Breaker Tripped Count', 'number of events',
- 'fielddata cache', 'es.evictions_tripped', 'line'],
+ 'fielddata cache', 'elastic.fielddata_evictions_tripped', 'line'],
'lines': [
- ['evictions', None, 'incremental'],
- ['tripped', None, 'incremental']
+ ['indices_fielddata_evictions', 'evictions', 'incremental'],
+ ['indices_fielddata_tripped', 'tripped', 'incremental']
]},
'cluster_health_nodes': {
'options': [None, 'Nodes And Tasks Statistics', 'units', 'cluster health API',
- 'es.cluster_health_nodes', 'stacked'],
+ 'elastic.cluster_health_nodes', 'stacked'],
'lines': [
- ['health_number_of_nodes', 'nodes', 'absolute'],
- ['health_number_of_data_nodes', 'data_nodes', 'absolute'],
- ['health_number_of_pending_tasks', 'pending_tasks', 'absolute'],
- ['health_number_of_in_flight_fetch', 'in_flight_fetch', 'absolute']
+ ['number_of_nodes', 'nodes', 'absolute'],
+ ['number_of_data_nodes', 'data_nodes', 'absolute'],
+ ['number_of_pending_tasks', 'pending_tasks', 'absolute'],
+ ['number_of_in_flight_fetch', 'in_flight_fetch', 'absolute']
]},
'cluster_health_status': {
'options': [None, 'Cluster Status', 'status', 'cluster health API',
- 'es.cluster_health_status', 'area'],
+ 'elastic.cluster_health_status', 'area'],
'lines': [
['status_green', 'green', 'absolute'],
['status_red', 'red', 'absolute'],
@@ -236,67 +239,67 @@ CHARTS = {
]},
'cluster_health_shards': {
'options': [None, 'Shards Statistics', 'shards', 'cluster health API',
- 'es.cluster_health_shards', 'stacked'],
+ 'elastic.cluster_health_shards', 'stacked'],
'lines': [
- ['health_active_shards', 'active_shards', 'absolute'],
- ['health_relocating_shards', 'relocating_shards', 'absolute'],
- ['health_unassigned_shards', 'unassigned', 'absolute'],
- ['health_delayed_unassigned_shards', 'delayed_unassigned', 'absolute'],
- ['health_initializing_shards', 'initializing', 'absolute'],
- ['health_active_shards_percent_as_number', 'active_percent', 'absolute']
+ ['active_shards', 'active_shards', 'absolute'],
+ ['relocating_shards', 'relocating_shards', 'absolute'],
+ ['unassigned_shards', 'unassigned', 'absolute'],
+ ['delayed_unassigned_shards', 'delayed_unassigned', 'absolute'],
+ ['initializing_shards', 'initializing', 'absolute'],
+ ['active_shards_percent_as_number', 'active_percent', 'absolute']
]},
'cluster_stats_nodes': {
'options': [None, 'Nodes Statistics', 'nodes', 'cluster stats API',
- 'es.cluster_nodes', 'stacked'],
+ 'elastic.cluster_nodes', 'stacked'],
'lines': [
- ['count_data_only', 'data_only', 'absolute'],
- ['count_master_data', 'master_data', 'absolute'],
- ['count_total', 'total', 'absolute'],
- ['count_master_only', 'master_only', 'absolute'],
- ['count_client', 'client', 'absolute']
+ ['nodes_count_data_only', 'data_only', 'absolute'],
+ ['nodes_count_master_data', 'master_data', 'absolute'],
+ ['nodes_count_total', 'total', 'absolute'],
+ ['nodes_count_master_only', 'master_only', 'absolute'],
+ ['nodes_count_client', 'client', 'absolute']
]},
'cluster_stats_query_cache': {
'options': [None, 'Query Cache Statistics', 'queries', 'cluster stats API',
- 'es.cluster_query_cache', 'stacked'],
+ 'elastic.cluster_query_cache', 'stacked'],
'lines': [
- ['query_cache_hit_count', 'hit', 'incremental'],
- ['query_cache_miss_count', 'miss', 'incremental']
+ ['indices_query_cache_hit_count', 'hit', 'incremental'],
+ ['indices_query_cache_miss_count', 'miss', 'incremental']
]},
'cluster_stats_docs': {
'options': [None, 'Docs Statistics', 'count', 'cluster stats API',
- 'es.cluster_docs', 'line'],
+ 'elastic.cluster_docs', 'line'],
'lines': [
- ['docs_count', 'docs', 'absolute']
+ ['indices_docs_count', 'docs', 'absolute']
]},
'cluster_stats_store': {
'options': [None, 'Store Statistics', 'MB', 'cluster stats API',
- 'es.cluster_store', 'line'],
+ 'elastic.cluster_store', 'line'],
'lines': [
- ['store_size_in_bytes', 'size', 'absolute', 1, 1048567]
+ ['indices_store_size_in_bytes', 'size', 'absolute', 1, 1048567]
]},
'cluster_stats_indices_shards': {
'options': [None, 'Indices And Shards Statistics', 'count', 'cluster stats API',
- 'es.cluster_indices_shards', 'stacked'],
+ 'elastic.cluster_indices_shards', 'stacked'],
'lines': [
['indices_count', 'indices', 'absolute'],
- ['shards_total', 'shards', 'absolute']
+ ['indices_shards_total', 'shards', 'absolute']
]},
'host_metrics_transport': {
'options': [None, 'Cluster Communication Transport Metrics', 'kilobit/s', 'host metrics',
- 'es.host_transport', 'area'],
+ 'elastic.host_transport', 'area'],
'lines': [
['transport_rx_size_in_bytes', 'in', 'incremental', 8, 1000],
['transport_tx_size_in_bytes', 'out', 'incremental', -8, 1000]
]},
'host_metrics_file_descriptors': {
'options': [None, 'Available File Descriptors In Percent', 'percent', 'host metrics',
- 'es.host_descriptors', 'area'],
+ 'elastic.host_descriptors', 'area'],
'lines': [
['file_descriptors_used', 'used', 'absolute', 1, 10]
]},
'host_metrics_http': {
'options': [None, 'Opened HTTP Connections', 'connections', 'host metrics',
- 'es.host_http_connections', 'line'],
+ 'elastic.host_http_connections', 'line'],
'lines': [
['http_current_open', 'opened', 'absolute', 1, 1]
]}
@@ -310,7 +313,9 @@ class Service(UrlService):
self.definitions = CHARTS
self.host = self.configuration.get('host')
self.port = self.configuration.get('port', 9200)
- self.scheme = self.configuration.get('scheme', 'http')
+ self.url = '{scheme}://{host}:{port}'.format(scheme=self.configuration.get('scheme', 'http'),
+ host=self.host,
+ port=self.port)
self.latency = dict()
self.methods = list()
@@ -329,33 +334,19 @@ class Service(UrlService):
self.error(str(error))
return False
- scheme = 'http' if self.scheme == 'http' else 'https'
- # Add handlers (auth, self signed cert accept)
- self.url = '%s://%s:%s' % (scheme, self.host, self.port)
- self.opener = self._build_opener()
# Create URL for every Elasticsearch API
- url_node_stats = '%s://%s:%s/_nodes/_local/stats' % (scheme, self.host, self.port)
- url_cluster_health = '%s://%s:%s/_cluster/health' % (scheme, self.host, self.port)
- url_cluster_stats = '%s://%s:%s/_cluster/stats' % (scheme, self.host, self.port)
-
- user_choice = [bool(self.configuration.get('node_stats', True)),
- bool(self.configuration.get('cluster_health', True)),
- bool(self.configuration.get('cluster_stats', True))]
-
- avail_methods = [METHODS(get_data_function=self._get_node_stats_,
- url=url_node_stats),
- METHODS(get_data_function=self._get_cluster_health_,
- url=url_cluster_health),
- METHODS(get_data_function=self._get_cluster_stats_,
- url=url_cluster_stats)]
+ self.methods = [METHODS(get_data=self._get_node_stats,
+ url=self.url + '/_nodes/_local/stats',
+ run=self.configuration.get('node_stats', True)),
+ METHODS(get_data=self._get_cluster_health,
+ url=self.url + '/_cluster/health',
+ run=self.configuration.get('cluster_health', True)),
+ METHODS(get_data=self._get_cluster_stats,
+ url=self.url + '/_cluster/stats',
+ run=self.configuration.get('cluster_stats', True))]
# Remove disabled API calls from 'avail methods'
- self.methods = [avail_methods[e[0]] for e in enumerate(avail_methods) if user_choice[e[0]]]
- data = self._get_data()
- if not data:
- return False
- self._data_from_check = data
- return True
+ return UrlService.check(self)
def _get_data(self):
threads = list()
@@ -363,7 +354,9 @@ class Service(UrlService):
result = dict()
for method in self.methods:
- th = Thread(target=method.get_data_function,
+ if not method.run:
+ continue
+ th = Thread(target=method.get_data,
args=(queue, method.url))
th.start()
threads.append(th)
@@ -374,7 +367,7 @@ class Service(UrlService):
return result or None
- def _get_cluster_health_(self, queue, url):
+ def _get_cluster_health(self, queue, url):
"""
Format data received from http request
:return: dict
@@ -387,7 +380,7 @@ class Service(UrlService):
data = loads(raw_data)
to_netdata = fetch_data_(raw_data=data,
- metrics_list=HEALTH_STATS)
+ metrics=HEALTH_STATS)
to_netdata.update({'status_green': 0, 'status_red': 0, 'status_yellow': 0,
'status_foo1': 0, 'status_foo2': 0, 'status_foo3': 0})
@@ -396,7 +389,7 @@ class Service(UrlService):
return queue.put(to_netdata)
- def _get_cluster_stats_(self, queue, url):
+ def _get_cluster_stats(self, queue, url):
"""
Format data received from http request
:return: dict
@@ -409,11 +402,11 @@ class Service(UrlService):
data = loads(raw_data)
to_netdata = fetch_data_(raw_data=data,
- metrics_list=CLUSTER_STATS)
+ metrics=CLUSTER_STATS)
return queue.put(to_netdata)
- def _get_node_stats_(self, queue, url):
+ def _get_node_stats(self, queue, url):
"""
Format data received from http request
:return: dict
@@ -428,23 +421,23 @@ class Service(UrlService):
node = list(data['nodes'].keys())[0]
to_netdata = fetch_data_(raw_data=data['nodes'][node],
- metrics_list=NODE_STATS)
+ metrics=NODE_STATS)
# Search, index, flush, fetch performance latency
for key in LATENCY:
try:
- to_netdata[key] = self.find_avg_(total=to_netdata[LATENCY[key]['total']],
- spent_time=to_netdata[LATENCY[key]['spent_time']],
- key=key)
+ to_netdata[key] = self.find_avg(total=to_netdata[LATENCY[key]['total']],
+ spent_time=to_netdata[LATENCY[key]['spent_time']],
+ key=key)
except KeyError:
continue
- if 'open_file_descriptors' in to_netdata and 'max_file_descriptors' in to_netdata:
- to_netdata['file_descriptors_used'] = round(float(to_netdata['open_file_descriptors'])
- / to_netdata['max_file_descriptors'] * 1000)
+ if 'process_open_file_descriptors' in to_netdata and 'process_max_file_descriptors' in to_netdata:
+ to_netdata['file_descriptors_used'] = round(float(to_netdata['process_open_file_descriptors'])
+ / to_netdata['process_max_file_descriptors'] * 1000)
return queue.put(to_netdata)
- def find_avg_(self, total, spent_time, key):
+ def find_avg(self, total, spent_time, key):
if key not in self.latency:
self.latency[key] = dict(total=total,
spent_time=spent_time)
@@ -459,17 +452,15 @@ class Service(UrlService):
return 0
-def fetch_data_(raw_data, metrics_list):
- to_netdata = dict()
- for metric, new_name, func in metrics_list:
+def fetch_data_(raw_data, metrics):
+ data = dict()
+ for metric in metrics:
value = raw_data
- for key in metric.split('.'):
- try:
- value = value[key]
- except KeyError:
- break
- if not isinstance(value, dict) and key:
- to_netdata[new_name or key] = value if not func else func(value)
-
- return to_netdata
-
+ metrics_list = metric.split('.')
+ try:
+ for m in metrics_list:
+ value = value[m]
+ except KeyError:
+ continue
+ data['_'.join(metrics_list)] = value
+ return data