1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
|
# -*- coding: utf-8 -*-
# Description: IPFS netdata python.d module
# Authors: Pawel Krupa (paulfantom), davidak
import json
from bases.FrameworkServices.UrlService import UrlService
# default module values (can be overridden per job in `config`)
# update_every = 2
priority = 60000
retries = 60
# default job configuration (overridden by python.d.plugin)
# config = {'local': {
# 'update_every': update_every,
# 'retries': retries,
# 'priority': priority,
# 'url': 'http://localhost:5001'
# }}
# charts order (can be overridden if you want less charts, or different order)
ORDER = ['bandwidth', 'peers', 'repo_size', 'repo_objects']
CHARTS = {
'bandwidth': {
'options': [None, 'IPFS Bandwidth', 'kbits/s', 'Bandwidth', 'ipfs.bandwidth', 'line'],
'lines': [
["in", None, "absolute", 8, 1000],
["out", None, "absolute", -8, 1000]
]},
'peers': {
'options': [None, 'IPFS Peers', 'peers', 'Peers', 'ipfs.peers', 'line'],
'lines': [
["peers", None, 'absolute']
]},
'repo_size': {
'options': [None, 'IPFS Repo Size', 'GB', 'Size', 'ipfs.repo_size', 'area'],
'lines': [
["avail", None, "absolute", 1, 1e9],
["size", None, "absolute", 1, 1e9],
]},
'repo_objects': {
'options': [None, 'IPFS Repo Objects', 'objects', 'Objects', 'ipfs.repo_objects', 'line'],
'lines': [
["objects", None, "absolute", 1, 1],
["pinned", None, "absolute", 1, 1],
["recursive_pins", None, "absolute", 1, 1]
]},
}
SI_zeroes = {'k': 3, 'm': 6, 'g': 9, 't': 12,
'p': 15, 'e': 18, 'z': 21, 'y': 24}
class Service(UrlService):
def __init__(self, configuration=None, name=None):
UrlService.__init__(self, configuration=configuration, name=name)
self.baseurl = self.configuration.get('url', 'http://localhost:5001')
self.order = ORDER
self.definitions = CHARTS
self.__storage_max = None
def _get_json(self, sub_url):
"""
:return: json decoding of the specified url
"""
self.url = self.baseurl + sub_url
try:
return json.loads(self._get_raw_data())
except (TypeError, ValueError):
return dict()
@staticmethod
def _recursive_pins(keys):
return len([k for k in keys if keys[k]["Type"] == b"recursive"])
@staticmethod
def _dehumanize(store_max):
# convert from '10Gb' to 10000000000
if not isinstance(store_max, int):
store_max = store_max.lower()
if store_max.endswith('b'):
val, units = store_max[:-2], store_max[-2]
if units in SI_zeroes:
val += '0'*SI_zeroes[units]
store_max = val
try:
store_max = int(store_max)
except (TypeError, ValueError):
store_max = None
return store_max
def _storagemax(self, store_cfg):
if self.__storage_max is None:
self.__storage_max = self._dehumanize(store_cfg['StorageMax'])
return self.__storage_max
def _get_data(self):
"""
Get data from API
:return: dict
"""
# suburl : List of (result-key, original-key, transform-func)
cfg = {
'/api/v0/stats/bw':
[('in', 'RateIn', int), ('out', 'RateOut', int)],
'/api/v0/swarm/peers':
[('peers', 'Strings', len)],
'/api/v0/stats/repo':
[('size', 'RepoSize', int), ('objects', 'NumObjects', int)],
'/api/v0/pin/ls':
[('pinned', 'Keys', len), ('recursive_pins', 'Keys', self._recursive_pins)],
'/api/v0/config/show': [('avail', 'Datastore', self._storagemax)]
}
r = dict()
for suburl in cfg:
in_json = self._get_json(suburl)
for new_key, orig_key, xmute in cfg[suburl]:
try:
r[new_key] = xmute(in_json[orig_key])
except Exception:
continue
return r or None
|