summaryrefslogtreecommitdiffstats
path: root/conf.d
diff options
context:
space:
mode:
authorFederico Ceratto <federico.ceratto@gmail.com>2017-12-19 23:39:21 +0000
committerFederico Ceratto <federico.ceratto@gmail.com>2017-12-19 23:39:21 +0000
commit61aedf201c2c4bf0e5aa4db32e74f4d860b88593 (patch)
treebcf4f9a0cd8bc2daf38b2ff9f29bfcc1e5ed8968 /conf.d
parentNew upstream version 1.8.0+dfsg (diff)
downloadnetdata-61aedf201c2c4bf0e5aa4db32e74f4d860b88593.tar.xz
netdata-61aedf201c2c4bf0e5aa4db32e74f4d860b88593.zip
New upstream version 1.9.0+dfsgupstream/1.9.0+dfsg
Diffstat (limited to '')
-rw-r--r--conf.d/Makefile.am11
-rw-r--r--conf.d/Makefile.in11
-rw-r--r--conf.d/apps_groups.conf18
-rw-r--r--conf.d/health.d/beanstalkd.conf36
-rw-r--r--conf.d/health.d/bind_rndc.conf2
-rw-r--r--conf.d/health.d/couchdb.conf13
-rw-r--r--conf.d/health.d/redis.conf20
-rw-r--r--conf.d/health.d/tcp_conn.conf19
-rw-r--r--conf.d/health.d/tcp_listen.conf27
-rw-r--r--conf.d/health.d/tcp_mem.conf20
-rw-r--r--conf.d/health.d/tcp_orphans.conf21
-rw-r--r--conf.d/health.d/tcp_resets.conf4
-rw-r--r--conf.d/health_alarm_notify.conf66
-rw-r--r--conf.d/python.d.conf14
-rw-r--r--conf.d/python.d/apache.conf21
-rw-r--r--conf.d/python.d/beanstalk.conf80
-rw-r--r--conf.d/python.d/bind_rndc.conf29
-rw-r--r--conf.d/python.d/chrony.conf21
-rw-r--r--conf.d/python.d/couchdb.conf91
-rw-r--r--conf.d/python.d/cpufreq.conf8
-rw-r--r--conf.d/python.d/dns_query_time.conf39
-rw-r--r--conf.d/python.d/dnsdist.conf85
-rw-r--r--conf.d/python.d/dovecot.conf21
-rw-r--r--conf.d/python.d/elasticsearch.conf21
-rw-r--r--conf.d/python.d/example.conf21
-rw-r--r--conf.d/python.d/exim.conf21
-rw-r--r--conf.d/python.d/fail2ban.conf29
-rw-r--r--conf.d/python.d/freeradius.conf42
-rw-r--r--conf.d/python.d/go_expvar.conf24
-rw-r--r--conf.d/python.d/haproxy.conf21
-rw-r--r--conf.d/python.d/hddtemp.conf24
-rw-r--r--conf.d/python.d/ipfs.conf21
-rw-r--r--conf.d/python.d/isc_dhcpd.conf32
-rw-r--r--conf.d/python.d/mdstat.conf8
-rw-r--r--conf.d/python.d/memcached.conf21
-rw-r--r--conf.d/python.d/mongodb.conf21
-rw-r--r--conf.d/python.d/mysql.conf21
-rw-r--r--conf.d/python.d/nginx.conf24
-rw-r--r--conf.d/python.d/nsd.conf21
-rw-r--r--conf.d/python.d/ovpn_status_log.conf23
-rw-r--r--conf.d/python.d/phpfpm.conf21
-rw-r--r--conf.d/python.d/postfix.conf21
-rw-r--r--conf.d/python.d/postgres.conf23
-rw-r--r--conf.d/python.d/powerdns.conf (renamed from conf.d/python.d/apache_cache.conf)46
-rw-r--r--conf.d/python.d/rabbitmq.conf21
-rw-r--r--conf.d/python.d/redis.conf21
-rw-r--r--conf.d/python.d/retroshare.conf21
-rw-r--r--conf.d/python.d/samba.conf24
-rw-r--r--conf.d/python.d/sensors.conf11
-rw-r--r--conf.d/python.d/smartd_log.conf33
-rw-r--r--conf.d/python.d/squid.conf21
-rw-r--r--conf.d/python.d/tomcat.conf21
-rw-r--r--conf.d/python.d/varnish.conf23
-rw-r--r--conf.d/python.d/web_log.conf21
-rw-r--r--conf.d/stream.conf35
55 files changed, 1077 insertions, 358 deletions
diff --git a/conf.d/Makefile.am b/conf.d/Makefile.am
index 7a0786678..095e891af 100644
--- a/conf.d/Makefile.am
+++ b/conf.d/Makefile.am
@@ -27,11 +27,13 @@ dist_nodeconfig_DATA = \
pythonconfigdir=$(configdir)/python.d
dist_pythonconfig_DATA = \
python.d/apache.conf \
- python.d/apache_cache.conf \
+ python.d/beanstalk.conf \
python.d/bind_rndc.conf \
python.d/chrony.conf \
+ python.d/couchdb.conf \
python.d/cpufreq.conf \
python.d/dns_query_time.conf \
+ python.d/dnsdist.conf \
python.d/dovecot.conf \
python.d/elasticsearch.conf \
python.d/example.conf \
@@ -53,6 +55,7 @@ dist_pythonconfig_DATA = \
python.d/phpfpm.conf \
python.d/postfix.conf \
python.d/postgres.conf \
+ python.d/powerdns.conf \
python.d/rabbitmq.conf \
python.d/redis.conf \
python.d/retroshare.conf \
@@ -70,8 +73,10 @@ healthconfigdir=$(configdir)/health.d
dist_healthconfig_DATA = \
health.d/apache.conf \
health.d/backend.conf \
+ health.d/beanstalkd.conf \
health.d/bind_rndc.conf \
health.d/cpu.conf \
+ health.d/couchdb.conf \
health.d/disks.conf \
health.d/elasticsearch.conf \
health.d/entropy.conf \
@@ -99,6 +104,10 @@ dist_healthconfig_DATA = \
health.d/softnet.conf \
health.d/squid.conf \
health.d/swap.conf \
+ health.d/tcp_conn.conf \
+ health.d/tcp_listen.conf \
+ health.d/tcp_mem.conf \
+ health.d/tcp_orphans.conf \
health.d/tcp_resets.conf \
health.d/udp_errors.conf \
health.d/varnish.conf \
diff --git a/conf.d/Makefile.in b/conf.d/Makefile.in
index 3d7084ed1..c1c291bcc 100644
--- a/conf.d/Makefile.in
+++ b/conf.d/Makefile.in
@@ -326,11 +326,13 @@ dist_nodeconfig_DATA = \
pythonconfigdir = $(configdir)/python.d
dist_pythonconfig_DATA = \
python.d/apache.conf \
- python.d/apache_cache.conf \
+ python.d/beanstalk.conf \
python.d/bind_rndc.conf \
python.d/chrony.conf \
+ python.d/couchdb.conf \
python.d/cpufreq.conf \
python.d/dns_query_time.conf \
+ python.d/dnsdist.conf \
python.d/dovecot.conf \
python.d/elasticsearch.conf \
python.d/example.conf \
@@ -352,6 +354,7 @@ dist_pythonconfig_DATA = \
python.d/phpfpm.conf \
python.d/postfix.conf \
python.d/postgres.conf \
+ python.d/powerdns.conf \
python.d/rabbitmq.conf \
python.d/redis.conf \
python.d/retroshare.conf \
@@ -368,8 +371,10 @@ healthconfigdir = $(configdir)/health.d
dist_healthconfig_DATA = \
health.d/apache.conf \
health.d/backend.conf \
+ health.d/beanstalkd.conf \
health.d/bind_rndc.conf \
health.d/cpu.conf \
+ health.d/couchdb.conf \
health.d/disks.conf \
health.d/elasticsearch.conf \
health.d/entropy.conf \
@@ -397,6 +402,10 @@ dist_healthconfig_DATA = \
health.d/softnet.conf \
health.d/squid.conf \
health.d/swap.conf \
+ health.d/tcp_conn.conf \
+ health.d/tcp_listen.conf \
+ health.d/tcp_mem.conf \
+ health.d/tcp_orphans.conf \
health.d/tcp_resets.conf \
health.d/udp_errors.conf \
health.d/varnish.conf \
diff --git a/conf.d/apps_groups.conf b/conf.d/apps_groups.conf
index 43bf1352c..4356e4910 100644
--- a/conf.d/apps_groups.conf
+++ b/conf.d/apps_groups.conf
@@ -100,14 +100,14 @@ puma: *puma*
# -----------------------------------------------------------------------------
# database servers
-sql: mysqld* mariad* postgres* oracle_* ora_*
-nosql: mongod redis* memcached
+sql: mysqld* mariad* postgres* postmaster* oracle_* ora_*
+nosql: mongod redis* memcached *couchdb*
timedb: prometheus *carbon-cache.py* *carbon-aggregator.py* *graphite/manage.py* *net.opentsdb.tools.TSDMain*
# -----------------------------------------------------------------------------
# email servers
-email: dovecot imapd pop3d amavis* master zmstat* zmmailboxdmgr qmgr oqmgr
+email: dovecot imapd pop3d amavis* master zmstat* zmmailboxdmgr qmgr oqmgr saslauthd opendkim clamd freshclam unbound tlsmgr postfwd2 postscreen postfix smtp* lmtp*
# -----------------------------------------------------------------------------
# network, routing, VPN
@@ -138,14 +138,15 @@ chat: irssi *vines* *prosody* murmurd
# -----------------------------------------------------------------------------
# monitoring
-logs: ulogd* syslog* rsyslog* logrotate systemd-journald
+logs: ulogd* syslog* rsyslog* logrotate systemd-journald rotatelogs
nms: snmpd vnstatd smokeping zabbix* monit munin* mon openhpid watchdog tailon nrpe
splunk: splunkd
-azure: mdsd *waagent* *omiserver* *omiagent* hv_kvp_daemon hv_vss_daemon
+azure: mdsd *waagent* *omiserver* *omiagent* hv_kvp_daemon hv_vss_daemon *auoms* *omsagent*
# -----------------------------------------------------------------------------
-# file systems and file servers
+# storage, file systems and file servers
+ceph: ceph-mds ceph-mgr ceph-mon ceph-osd radosgw* rbd-*
samba: smbd nmbd winbindd
nfs: rpcbind rpc.* nfs*
zfs: spl_* z_* txg_* zil_* arc_* l2arc*
@@ -182,11 +183,12 @@ dhcp: *dhcp*
# name servers and clients
named: named rncd dig
+dnsdist: dnsdist
# -----------------------------------------------------------------------------
# installation / compilation / debugging
-build: cc1 cc1plus as gcc* cppcheck ld make cmake automake autoconf autoreconf
+build: cc1 cc1plus as gcc* cppcheck ld make cmake automake autoconf autoreconf
build: git gdb valgrind*
# -----------------------------------------------------------------------------
@@ -273,6 +275,8 @@ kernel: fsnotify_mark kthrotld deferwq scsi_*
kafka: *kafka.Kafka*
+rabbitmq: *rabbitmq*
+
sidekiq: *sidekiq*
java: java
ipfs: ipfs
diff --git a/conf.d/health.d/beanstalkd.conf b/conf.d/health.d/beanstalkd.conf
new file mode 100644
index 000000000..30dc27328
--- /dev/null
+++ b/conf.d/health.d/beanstalkd.conf
@@ -0,0 +1,36 @@
+# get the number of buried jobs in all queues
+
+template: server_buried_jobs
+ on: beanstalk.current_jobs
+ calc: $buried
+ units: jobs
+ every: 10s
+ warn: $this > 0
+ crit: $this > 10
+ delay: up 0 down 5m multiplier 1.2 max 1h
+ info: the number of buried jobs aggregated across all tubes
+ to: sysadmin
+
+# get the number of buried jobs per queue
+
+#template: tube_buried_jobs
+# on: beanstalk.jobs
+# calc: $buried
+# units: jobs
+# every: 10s
+# warn: $this > 0
+# crit: $this > 10
+# delay: up 0 down 5m multiplier 1.2 max 1h
+# info: the number of jobs buried per tube
+# to: sysadmin
+
+# get the current number of tubes
+
+#template: number_of_tubes
+# on: beanstalk.current_tubes
+# calc: $tubes
+# every: 10s
+# warn: $this < 5
+# delay: up 0 down 5m multiplier 1.2 max 1h
+# info: the current number of tubes on the server
+# to: sysadmin
diff --git a/conf.d/health.d/bind_rndc.conf b/conf.d/health.d/bind_rndc.conf
index 028bc9d08..4145e77cd 100644
--- a/conf.d/health.d/bind_rndc.conf
+++ b/conf.d/health.d/bind_rndc.conf
@@ -1,4 +1,4 @@
- alarm: bind_rndc_stats_file_size
+ template: bind_rndc_stats_file_size
on: bind_rndc.stats_size
units: megabytes
every: 60
diff --git a/conf.d/health.d/couchdb.conf b/conf.d/health.d/couchdb.conf
new file mode 100644
index 000000000..4a2895280
--- /dev/null
+++ b/conf.d/health.d/couchdb.conf
@@ -0,0 +1,13 @@
+
+# make sure couchdb is running
+
+template: couchdb_last_collected_secs
+ on: couchdb.request_methods
+ calc: $now - $last_collected_t
+ units: seconds ago
+ every: 10s
+ warn: $this > (($status >= $WARNING) ? ($update_every) : ( 5 * $update_every))
+ crit: $this > (($status == $CRITICAL) ? ($update_every) : (60 * $update_every))
+ delay: down 5m multiplier 1.5 max 1h
+ info: number of seconds since the last successful data collection
+ to: dba
diff --git a/conf.d/health.d/redis.conf b/conf.d/health.d/redis.conf
index 5f6d397ea..c08a884a6 100644
--- a/conf.d/health.d/redis.conf
+++ b/conf.d/health.d/redis.conf
@@ -12,3 +12,23 @@ template: redis_last_collected_secs
info: number of seconds since the last successful data collection
to: dba
+template: redis_bgsave_broken
+families: *
+ on: redis.bgsave_health
+ every: 10s
+ crit: $rdb_last_bgsave_status != 0
+ units: ok/failed
+ info: states if redis bgsave is working
+ delay: down 5m multiplier 1.5 max 1h
+ to: dba
+
+template: redis_bgsave_slow
+families: *
+ on: redis.bgsave_now
+ every: 10s
+ warn: $rdb_bgsave_in_progress > 600
+ crit: $rdb_bgsave_in_progress > 1200
+ units: seconds
+ info: the time redis needs to save its database
+ delay: down 5m multiplier 1.5 max 1h
+ to: dba
diff --git a/conf.d/health.d/tcp_conn.conf b/conf.d/health.d/tcp_conn.conf
new file mode 100644
index 000000000..7aa9a9800
--- /dev/null
+++ b/conf.d/health.d/tcp_conn.conf
@@ -0,0 +1,19 @@
+
+#
+# ${tcp_max_connections} may be nan or -1 if the system
+# supports dynamic threshold for TCP connections.
+# In this case, the alarm will always be zero.
+#
+
+ alarm: tcp_connections
+ on: ipv4.tcpsock
+ os: linux
+ hosts: *
+ calc: (${tcp_max_connections} > 0) ? ( ${connections} * 100 / ${tcp_max_connections} ) : 0
+ units: %
+ every: 10s
+ warn: $this > (($status >= $WARNING ) ? ( 60 ) : ( 80 ))
+ crit: $this > (($status >= $CRITICAL) ? ( 80 ) : ( 90 ))
+ delay: up 0 down 5m multiplier 1.5 max 1h
+ info: the percentage of IPv4 TCP connections over the max allowed
+ to: sysadmin
diff --git a/conf.d/health.d/tcp_listen.conf b/conf.d/health.d/tcp_listen.conf
new file mode 100644
index 000000000..957964ae4
--- /dev/null
+++ b/conf.d/health.d/tcp_listen.conf
@@ -0,0 +1,27 @@
+# -----------------------------------------------------------------------------
+# tcp listen sockets issues
+
+ alarm: 1m_ipv4_tcp_listen_overflows
+ on: ipv4.tcplistenissues
+ os: linux freebsd
+ hosts: *
+ lookup: sum -60s unaligned absolute of ListenOverflows
+ units: overflows
+ every: 10s
+ crit: $this > 0
+ delay: up 0 down 5m multiplier 1.5 max 1h
+ info: the number of TCP listen socket overflows during the last minute
+ to: sysadmin
+
+ alarm: 1m_ipv4_tcp_listen_drops
+ on: ipv4.tcplistenissues
+ os: linux
+ hosts: *
+ lookup: sum -60s unaligned absolute of ListenDrops
+ units: drops
+ every: 10s
+ crit: $this > 0
+ delay: up 0 down 5m multiplier 1.5 max 1h
+ info: the number of TCP listen socket drops during the last minute
+ to: sysadmin
+
diff --git a/conf.d/health.d/tcp_mem.conf b/conf.d/health.d/tcp_mem.conf
new file mode 100644
index 000000000..6927d5765
--- /dev/null
+++ b/conf.d/health.d/tcp_mem.conf
@@ -0,0 +1,20 @@
+#
+# check
+# http://blog.tsunanet.net/2011/03/out-of-socket-memory.html
+#
+# We give a warning when TCP is under memory pressure
+# and a critical when TCP is 90% of its upper memory limit
+#
+
+ alarm: tcp_memory
+ on: ipv4.sockstat_tcp_mem
+ os: linux
+ hosts: *
+ calc: ${mem} * 100 / ${tcp_mem_high}
+ units: %
+ every: 10s
+ warn: ${mem} > (($status >= $WARNING ) ? ( ${tcp_mem_pressure} * 0.8 ) : ( ${tcp_mem_pressure} ))
+ crit: ${mem} > (($status >= $CRITICAL ) ? ( ${tcp_mem_pressure} ) : ( ${tcp_mem_high} * 0.9 ))
+ delay: up 0 down 5m multiplier 1.5 max 1h
+ info: the amount of TCP memory as a percentage of its max memory limit
+ to: sysadmin
diff --git a/conf.d/health.d/tcp_orphans.conf b/conf.d/health.d/tcp_orphans.conf
new file mode 100644
index 000000000..280d6590f
--- /dev/null
+++ b/conf.d/health.d/tcp_orphans.conf
@@ -0,0 +1,21 @@
+
+#
+# check
+# http://blog.tsunanet.net/2011/03/out-of-socket-memory.html
+#
+# The kernel may penalize orphans by 2x or even 4x
+# so we alarm warning at 25% and critical at 50%
+#
+
+ alarm: tcp_orphans
+ on: ipv4.sockstat_tcp_sockets
+ os: linux
+ hosts: *
+ calc: ${orphan} * 100 / ${tcp_max_orphans}
+ units: %
+ every: 10s
+ warn: $this > (($status >= $WARNING ) ? ( 20 ) : ( 25 ))
+ crit: $this > (($status >= $CRITICAL) ? ( 25 ) : ( 50 ))
+ delay: up 0 down 5m multiplier 1.5 max 1h
+ info: the percentage of orphan IPv4 TCP sockets over the max allowed (this may lead to too-many-orphans errors)
+ to: sysadmin
diff --git a/conf.d/health.d/tcp_resets.conf b/conf.d/health.d/tcp_resets.conf
index fec124ac7..e6cfd39ab 100644
--- a/conf.d/health.d/tcp_resets.conf
+++ b/conf.d/health.d/tcp_resets.conf
@@ -37,7 +37,7 @@
every: 10s
warn: $this > ((($1m_ipv4_tcp_resets_sent < 5)?(5):($1m_ipv4_tcp_resets_sent)) * (($status >= $WARNING) ? (1) : (20)))
delay: up 0 down 60m multiplier 1.2 max 2h
-options: no-clear-notification
+ options: no-clear-notification
info: average TCP RESETS this host is sending, over the last 10 seconds (this can be an indication that a port scan is made, or that a service running on this host has crashed; clear notification for this alarm will not be sent)
to: sysadmin
@@ -62,6 +62,6 @@ options: no-clear-notification
every: 10s
warn: $this > ((($1m_ipv4_tcp_resets_received < 5)?(5):($1m_ipv4_tcp_resets_received)) * (($status >= $WARNING) ? (1) : (10)))
delay: up 0 down 60m multiplier 1.2 max 2h
-options: no-clear-notification
+ options: no-clear-notification
info: average TCP RESETS this host is receiving, over the last 10 seconds (this can be an indication that a service this host needs, has crashed; clear notification for this alarm will not be sent)
to: sysadmin
diff --git a/conf.d/health_alarm_notify.conf b/conf.d/health_alarm_notify.conf
index 641272ced..eb01e2bb9 100644
--- a/conf.d/health_alarm_notify.conf
+++ b/conf.d/health_alarm_notify.conf
@@ -7,6 +7,7 @@
# - e-mails (using the sendmail command),
# - push notifications to your mobile phone (pushover.net),
# - messages to your slack team (slack.com),
+# - messages to your flock team (flock.com),
# - messages to your discord guild (discordapp.com),
# - messages to your telegram chat / group chat (telegram.org)
# - sms messages to your cell phone or any sms enabled device (twilio.com)
@@ -23,7 +24,7 @@
# proxy configuration
#
# If you need to send curl based notifications (pushover, pushbullet, slack,
-# discord, telegram) via a proxy, set these to your proxy address:
+# flock, discord, telegram) via a proxy, set these to your proxy address:
#export http_proxy="http://10.0.0.1:3128/"
#export https_proxy="http://10.0.0.1:3128/"
@@ -63,6 +64,7 @@ curl=""
# - pushover user tokens
# - telegram chat ids
# - slack channels
+# - flock rooms
# - discord channels
# - hipchat rooms
# - sms phone numbers
@@ -77,9 +79,11 @@ curl=""
# pushover : "2987343...9437837 8756278...2362736|critical"
# telegram : "111827421 112746832|critical"
# slack : "alarms disasters|critical"
+# flock : "alarms disasters|critical"
# discord : "alarms disasters|critical"
# twilio : "+15555555555 +17777777777|critical"
# messagebird: "+15555555555 +17777777777|critical"
+# kavenegar : "09155555555 09177777777|critical"
# pd : "<pd_service_key_1> <pd_service_key_2>|critical"
#
# If a recipient is set to empty string, the default recipient of the given
@@ -160,6 +164,9 @@ SEND_PUSHBULLET="YES"
PUSHBULLET_ACCESS_TOKEN=""
DEFAULT_RECIPIENT_PUSHBULLET=""
+# Device iden of the sending device. Optional.
+PUSHBULLET_SOURCE_DEVICE=""
+
#------------------------------------------------------------------------------
# Twilio (twilio.com) SMS options
@@ -205,6 +212,28 @@ DEFAULT_RECIPIENT_MESSAGEBIRD=""
#------------------------------------------------------------------------------
+# Kavenegar (Kavenegar.com) SMS options
+
+# multiple recipients can be given like this:
+# "09155555555 09177777777"
+
+# enable/disable sending kavenegar SMS
+SEND_KAVENEGAR="YES"
+
+# to get an access key, after selecting and purchasing your desired service
+# at http://kavenegar.com/pricing.html
+# login to your account, go to your dashboard and my account are
+# https://panel.kavenegar.com/Client/setting/account from API Key
+# copy your api key. You can generate new API Key too.
+# You can find and select kevenegar sender number from this place.
+
+# Without an API key, netdata cannot send KAVENEGAR text messages.
+KAVENEGAR_API_KEY=""
+KAVENEGAR_SENDER=""
+DEFAULT_RECIPIENT_KAVENEGAR=""
+
+
+#------------------------------------------------------------------------------
# telegram (telegram.org) global notification options
# To get your chat ID send the command /my_id to telegram bot @get_id.
@@ -247,6 +276,21 @@ DEFAULT_RECIPIENT_SLACK=""
#------------------------------------------------------------------------------
+# flock (flock.com) global notification options
+
+# enable/disable sending flock notifications
+SEND_FLOCK="YES"
+
+# Login to flock.com and create an incoming webhook. You need only one for all
+# your netdata servers (or you can have one for each of your netdata).
+# Without it, netdata cannot send flock notifications.
+FLOCK_WEBHOOK_URL=""
+
+# if a role recipient is not configured, no notification will be sent
+DEFAULT_RECIPIENT_FLOCK=""
+
+
+#------------------------------------------------------------------------------
# discord (discordapp.com) global notification options
# multiple recipients can be given like this:
@@ -398,6 +442,8 @@ role_recipients_telegram[sysadmin]="${DEFAULT_RECIPIENT_TELEGRAM}"
role_recipients_slack[sysadmin]="${DEFAULT_RECIPIENT_SLACK}"
+role_recipients_flock[sysadmin]="${DEFAULT_RECIPIENT_FLOCK}"
+
role_recipients_discord[sysadmin]="${DEFAULT_RECIPIENT_DISCORD}"
role_recipients_hipchat[sysadmin]="${DEFAULT_RECIPIENT_HIPCHAT}"
@@ -406,6 +452,8 @@ role_recipients_twilio[sysadmin]="${DEFAULT_RECIPIENT_TWILIO}"
role_recipients_messagebird[sysadmin]="${DEFAULT_RECIPIENT_MESSAGEBIRD}"
+role_recipients_kavenegar[sysadmin]="${DEFAULT_RECIPIENT_KAVENEGAR}"
+
role_recipients_pd[sysadmin]="${DEFAULT_RECIPIENT_PD}"
role_recipients_custom[sysadmin]="${DEFAULT_RECIPIENT_CUSTOM}"
@@ -423,6 +471,8 @@ role_recipients_telegram[domainadmin]="${DEFAULT_RECIPIENT_TELEGRAM}"
role_recipients_slack[domainadmin]="${DEFAULT_RECIPIENT_SLACK}"
+role_recipients_flock[domainadmin]="${DEFAULT_RECIPIENT_FLOCK}"
+
role_recipients_discord[domainadmin]="${DEFAULT_RECIPIENT_DISCORD}"
role_recipients_hipchat[domainadmin]="${DEFAULT_RECIPIENT_HIPCHAT}"
@@ -431,6 +481,8 @@ role_recipients_twilio[domainadmin]="${DEFAULT_RECIPIENT_TWILIO}"
role_recipients_messagebird[domainadmin]="${DEFAULT_RECIPIENT_MESSAGEBIRD}"
+role_recipients_kavenegar[domainadmin]="${DEFAULT_RECIPIENT_KAVENEGAR}"
+
role_recipients_pd[domainadmin]="${DEFAULT_RECIPIENT_PD}"
role_recipients_custom[domainadmin]="${DEFAULT_RECIPIENT_CUSTOM}"
@@ -449,6 +501,8 @@ role_recipients_telegram[dba]="${DEFAULT_RECIPIENT_TELEGRAM}"
role_recipients_slack[dba]="${DEFAULT_RECIPIENT_SLACK}"
+role_recipients_flock[dba]="${DEFAULT_RECIPIENT_FLOCK}"
+
role_recipients_discord[dba]="${DEFAULT_RECIPIENT_DISCORD}"
role_recipients_hipchat[dba]="${DEFAULT_RECIPIENT_HIPCHAT}"
@@ -457,6 +511,8 @@ role_recipients_twilio[dba]="${DEFAULT_RECIPIENT_TWILIO}"
role_recipients_messagebird[dba]="${DEFAULT_RECIPIENT_MESSAGEBIRD}"
+role_recipients_kavenegar[dba]="${DEFAULT_RECIPIENT_KAVENEGAR}"
+
role_recipients_pd[dba]="${DEFAULT_RECIPIENT_PD}"
role_recipients_custom[dba]="${DEFAULT_RECIPIENT_CUSTOM}"
@@ -475,6 +531,8 @@ role_recipients_telegram[webmaster]="${DEFAULT_RECIPIENT_TELEGRAM}"
role_recipients_slack[webmaster]="${DEFAULT_RECIPIENT_SLACK}"
+role_recipients_flock[webmaster]="${DEFAULT_RECIPIENT_FLOCK}"
+
role_recipients_discord[webmaster]="${DEFAULT_RECIPIENT_DISCORD}"
role_recipients_hipchat[webmaster]="${DEFAULT_RECIPIENT_HIPCHAT}"
@@ -483,6 +541,8 @@ role_recipients_twilio[webmaster]="${DEFAULT_RECIPIENT_TWILIO}"
role_recipients_messagebird[webmaster]="${DEFAULT_RECIPIENT_MESSAGEBIRD}"
+role_recipients_kavenegar[webmaster]="${DEFAULT_RECIPIENT_KAVENEGAR}"
+
role_recipients_pd[webmaster]="${DEFAULT_RECIPIENT_PD}"
role_recipients_custom[webmaster]="${DEFAULT_RECIPIENT_CUSTOM}"
@@ -501,6 +561,8 @@ role_recipients_telegram[proxyadmin]="${DEFAULT_RECIPIENT_TELEGRAM}"
role_recipients_slack[proxyadmin]="${DEFAULT_RECIPIENT_SLACK}"
+role_recipients_flock[proxyadmin]="${DEFAULT_RECIPIENT_FLOCK}"
+
role_recipients_discord[proxyadmin]="${DEFAULT_RECIPIENT_DISCORD}"
role_recipients_hipchat[proxyadmin]="${DEFAULT_RECIPIENT_HIPCHAT}"
@@ -509,6 +571,8 @@ role_recipients_twilio[proxyadmin]="${DEFAULT_RECIPIENT_TWILIO}"
role_recipients_messagebird[proxyadmin]="${DEFAULT_RECIPIENT_MESSAGEBIRD}"
+role_recipients_kavenegar[proxyadmin]="${DEFAULT_RECIPIENT_KAVENEGAR}"
+
role_recipients_pd[proxyadmin]="${DEFAULT_RECIPIENT_PD}"
role_recipients_custom[proxyadmin]="${DEFAULT_RECIPIENT_CUSTOM}"
diff --git a/conf.d/python.d.conf b/conf.d/python.d.conf
index 741d49914..2c3d400ca 100644
--- a/conf.d/python.d.conf
+++ b/conf.d/python.d.conf
@@ -9,13 +9,6 @@
# Enable / disable the whole python.d.plugin (all its modules)
enabled: yes
-# Prevent log flood
-# Define how many log messages can be written to log file in one log_interval
-logs_per_interval: 200
-
-# Define how long is one logging interval (in seconds)
-log_interval: 3600
-
# ----------------------------------------------------------------------
# Enable / Disable python.d.plugin modules
#default_run: yes
@@ -29,11 +22,14 @@ log_interval: 3600
# apache_cache has been replaced by web_log
apache_cache: no
# apache: yes
+# beanstalk: yes
# bind_rndc: yes
-# chrony: yes
+chrony: no
+# couchdb: yes
# cpufreq: yes
# cpuidle: yes
# dns_query_time: yes
+# dnsdist: yes
# dovecot: yes
# elasticsearch: yes
@@ -53,6 +49,7 @@ go_expvar: no
# isc_dhcpd: yes
# mdstat: yes
# memcached: yes
+# mongodb: yes
# mysql: yes
# nginx: yes
# nsd: yes
@@ -64,6 +61,7 @@ nginx_log: no
# phpfpm: yes
# postfix: yes
# postgres: yes
+# powerdns: yes
# rabbitmq: yes
# redis: yes
# retroshare: yes
diff --git a/conf.d/python.d/apache.conf b/conf.d/python.d/apache.conf
index 5b151ef70..3bbc3f786 100644
--- a/conf.d/python.d/apache.conf
+++ b/conf.d/python.d/apache.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, apache also supports the following:
#
diff --git a/conf.d/python.d/beanstalk.conf b/conf.d/python.d/beanstalk.conf
new file mode 100644
index 000000000..940801877
--- /dev/null
+++ b/conf.d/python.d/beanstalk.conf
@@ -0,0 +1,80 @@
+# netdata python.d.plugin configuration for beanstalk
+#
+# This file is in YaML format. Generally the format is:
+#
+# name: value
+#
+# There are 2 sections:
+# - global variables
+# - one or more JOBS
+#
+# JOBS allow you to collect values from multiple sources.
+# Each source will have its own set of charts.
+#
+# JOB parameters have to be indented (using spaces only, example below).
+
+# ----------------------------------------------------------------------
+# Global Variables
+# These variables set the defaults for all JOBs, however each JOB
+# may define its own, overriding the defaults.
+
+# update_every sets the default data collection frequency.
+# If unset, the python.d.plugin default is used.
+# update_every: 1
+
+# priority controls the order of charts at the netdata dashboard.
+# Lower numbers move the charts towards the top of the page.
+# If unset, the default for python.d.plugin is used.
+# priority: 60000
+
+# retries sets the number of retries to be made in case of failures.
+# If unset, the default for python.d.plugin is used.
+# Attempts to restore the service are made once every update_every
+# and only if the module has collected values in the past.
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
+
+# chart_cleanup sets the default chart cleanup interval in iterations.
+# A chart is marked as obsolete if it has not been updated
+# 'chart_cleanup' iterations in a row.
+# When a plugin sends the obsolete flag, the charts are not deleted
+# from netdata immediately.
+# They will be hidden immediately (not offered to dashboard viewer,
+# streamed upstream and archived to backends) and deleted one hour
+# later (configurable from netdata.conf).
+# chart_cleanup: 10
+
+# ----------------------------------------------------------------------
+# JOBS (data collection sources)
+#
+# The default JOBS share the same *name*. JOBS with the same name
+# are mutually exclusive. Only one of them will be allowed running at
+# any time. This allows autodetection to try several alternatives and
+# pick the one that works.
+#
+# Any number of jobs is supported.
+#
+# All python.d.plugin JOBS (for all its modules) support a set of
+# predefined parameters. These are:
+#
+# job_name:
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
+# chart_cleanup: 10 # the JOB's chart cleanup interval in iterations
+#
+# Additionally to the above, apache also supports the following:
+#
+# host: 'host' # Server ip address or hostname. Default: 127.0.0.1
+# port: port # Beanstalkd port. Default:
+#
+# ----------------------------------------------------------------------
diff --git a/conf.d/python.d/bind_rndc.conf b/conf.d/python.d/bind_rndc.conf
index e4f7ac825..71958ff98 100644
--- a/conf.d/python.d/bind_rndc.conf
+++ b/conf.d/python.d/bind_rndc.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, bind_rndc also supports the following:
#
@@ -75,7 +82,7 @@
# chown bind bind/
#
# 4. RELOAD (NOT restart) BIND
-# systemctl reload bind9.serice
+# systemctl reload bind9.service
#
# 5. Run as a root 'rndc stats' to dump (BIND will create named.stats in new directory)
#
@@ -101,9 +108,5 @@
# To test your logrotate conf file run as root:
#
# logrotate /etc/logrotate.d/bind-rndc -d (debug dry-run mode)
-# ------------------------------------------------------------------------------------------------------------------
-# AUTO-DETECTION JOBS
-# only one of them will run (they have the same name)
#
-#local:
-# named_stats_path: '/var/log/bind/named.stats'
+# ----------------------------------------------------------------------
diff --git a/conf.d/python.d/chrony.conf b/conf.d/python.d/chrony.conf
index 46229687b..9ac906b5f 100644
--- a/conf.d/python.d/chrony.conf
+++ b/conf.d/python.d/chrony.conf
@@ -31,7 +31,13 @@ update_every: 5
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@ update_every: 5
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, chrony also supports the following:
#
diff --git a/conf.d/python.d/couchdb.conf b/conf.d/python.d/couchdb.conf
new file mode 100644
index 000000000..5f6e75cff
--- /dev/null
+++ b/conf.d/python.d/couchdb.conf
@@ -0,0 +1,91 @@
+# netdata python.d.plugin configuration for couchdb
+#
+# This file is in YaML format. Generally the format is:
+#
+# name: value
+#
+# There are 2 sections:
+# - global variables
+# - one or more JOBS
+#
+# JOBS allow you to collect values from multiple sources.
+# Each source will have its own set of charts.
+#
+# JOB parameters have to be indented (using spaces only, example below).
+
+# ----------------------------------------------------------------------
+# Global Variables
+# These variables set the defaults for all JOBs, however each JOB
+# may define its own, overriding the defaults.
+
+# update_every sets the default data collection frequency.
+# If unset, the python.d.plugin default is used.
+# By default, CouchDB only updates its stats every 10 seconds.
+update_every: 10
+
+# priority controls the order of charts at the netdata dashboard.
+# Lower numbers move the charts towards the top of the page.
+# If unset, the default for python.d.plugin is used.
+# priority: 60000
+
+# retries sets the number of retries to be made in case of failures.
+# If unset, the default for python.d.plugin is used.
+# Attempts to restore the service are made once every update_every
+# and only if the module has collected values in the past.
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
+
+# ----------------------------------------------------------------------
+# JOBS (data collection sources)
+#
+# The default JOBS share the same *name*. JOBS with the same name
+# are mutually exclusive. Only one of them will be allowed running at
+# any time. This allows autodetection to try several alternatives and
+# pick the one that works.
+#
+# Any number of jobs is supported.
+#
+# All python.d.plugin JOBS (for all its modules) support a set of
+# predefined parameters. These are:
+#
+# job_name:
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
+#
+# Additionally to the above, the couchdb plugin also supports the following:
+#
+# host: 'ipaddress' # Server ip address or hostname. Default: 127.0.0.1
+# port: 'port' # CouchDB port. Default: 15672
+# scheme: 'scheme' # http or https. Default: http
+# node: 'couchdb@127.0.0.1' # CouchDB node name. Same as -name vm.args argument.
+#
+# if the URL is password protected, the following are supported:
+#
+# user: 'username'
+# pass: 'password'
+#
+# if db-specific stats are desired, place their names in databases:
+# databases: 'npm-registry animaldb'
+#
+# ----------------------------------------------------------------------
+# AUTO-DETECTION JOBS
+# only one of them will run (they have the same name)
+#
+localhost:
+ name: 'local'
+ host: '127.0.0.1'
+ port: '5984'
+ node: 'couchdb@127.0.0.1'
+ scheme: 'http'
+# user: 'admin'
+# pass: 'password'
diff --git a/conf.d/python.d/cpufreq.conf b/conf.d/python.d/cpufreq.conf
index 10c96917f..0890245d9 100644
--- a/conf.d/python.d/cpufreq.conf
+++ b/conf.d/python.d/cpufreq.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# The directory to search for the file scaling_cur_freq
sys_dir: "/sys/devices"
diff --git a/conf.d/python.d/dns_query_time.conf b/conf.d/python.d/dns_query_time.conf
index f4d4dbf92..d32c6db83 100644
--- a/conf.d/python.d/dns_query_time.conf
+++ b/conf.d/python.d/dns_query_time.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,26 +53,19 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, dns_query_time also supports the following:
#
-# dns_servers: 'dns servers' # list of dns servers to query
-# domains: 'domains' # list of domains
-# aggregate: yes/no # Default: yes. Aggregate all servers in one chart or not
-# response_timeout: 4 # Defalt: 4. Dns query response timeout (query = -100 if response time > response_time)
-#
-#
-# ----------------------------------------------------------------------
-# AUTO-DETECTION JOBS
-# only one of them will run (they have the same name)
-#
+# dns_servers: 'dns servers' # List of dns servers to query
+# domains: 'domains' # List of domains
+# aggregate: yes/no # Aggregate all servers in one chart or not
+# response_timeout: 4 # Dns query response timeout (query = -100 if response time > response_time)
#
-#aggregate: yes
-#dns_servers: '8.8.8.8 8.8.4.4'
-#domains: 'python.org distrowatch.com linuxmint.com linux.com rsyslog.com liblognorm.com archlinux.org cisco.com debian.org kernel.org gns3.com opera.com github.com youtube.com amazon.co.uk kde.org netdata.firehol.org ubuntu.com redhat.com opensuse.org wireshark.org vmware.com microsoft.com elastic.co'
+# ---------------------------------------------------------------------- \ No newline at end of file
diff --git a/conf.d/python.d/dnsdist.conf b/conf.d/python.d/dnsdist.conf
new file mode 100644
index 000000000..aec58b8e1
--- /dev/null
+++ b/conf.d/python.d/dnsdist.conf
@@ -0,0 +1,85 @@
+# netdata python.d.plugin configuration for dnsdist
+#
+# This file is in YaML format. Generally the format is:
+#
+# name: value
+#
+# There are 2 sections:
+# - global variables
+# - one or more JOBS
+#
+# JOBS allow you to collect values from multiple sources.
+# Each source will have its own set of charts.
+#
+# JOB parameters have to be indented (using spaces only, example below).
+
+# ----------------------------------------------------------------------
+# Global Variables
+# These variables set the defaults for all JOBs, however each JOB
+# may define its own, overriding the defaults.
+
+# update_every sets the default data collection frequency.
+# If unset, the python.d.plugin default is used.
+#update_every: 1
+
+# priority controls the order of charts at the netdata dashboard.
+# Lower numbers move the charts towards the top of the page.
+# If unset, the default for python.d.plugin is used.
+# priority: 60000
+
+# retries sets the number of retries to be made in case of failures.
+# If unset, the default for python.d.plugin is used.
+# Attempts to restore the service are made once every update_every
+# and only if the module has collected values in the past.
+#retries: 600000
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+#autodetection_retry: 1
+
+# ----------------------------------------------------------------------
+# JOBS (data collection sources)
+#
+# The default JOBS share the same *name*. JOBS with the same name
+# are mutually exclusive. Only one of them will be allowed running at
+# any time. This allows autodetection to try several alternatives and
+# pick the one that works.
+#
+# Any number of jobs is supported.
+#
+# All python.d.plugin JOBS (for all its modules) support a set of
+# predefined parameters. These are:
+#
+# job_name:
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
+#
+#
+# Additionally to the above, dnsdist also supports the following:
+#
+# url: 'URL' # the URL to fetch dnsdist performance statistics
+# user: 'username' # username for basic auth
+# pass: 'password' # password for basic auth
+# header:
+# X-API-Key: 'Key' # API key
+#
+# ----------------------------------------------------------------------
+# AUTO-DETECTION JOBS
+# only one of them will run (they have the same name)
+
+# localhost:
+# name : 'local'
+# url : 'http://127.0.0.1:5053/jsonstat?command=stats'
+# user : 'username'
+# pass : 'password'
+# header:
+# X-API-Key: 'dnsdist-api-key'
+
+
diff --git a/conf.d/python.d/dovecot.conf b/conf.d/python.d/dovecot.conf
index 917c5272e..56c394991 100644
--- a/conf.d/python.d/dovecot.conf
+++ b/conf.d/python.d/dovecot.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, dovecot also supports the following:
#
diff --git a/conf.d/python.d/elasticsearch.conf b/conf.d/python.d/elasticsearch.conf
index 7c35df229..213843bf9 100644
--- a/conf.d/python.d/elasticsearch.conf
+++ b/conf.d/python.d/elasticsearch.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, elasticsearch plugin also supports the following:
#
diff --git a/conf.d/python.d/example.conf b/conf.d/python.d/example.conf
index 31f9a49a0..e7fed9b50 100644
--- a/conf.d/python.d/example.conf
+++ b/conf.d/python.d/example.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, example also supports the following:
#
diff --git a/conf.d/python.d/exim.conf b/conf.d/python.d/exim.conf
index 07d72c5a3..2add7b2cb 100644
--- a/conf.d/python.d/exim.conf
+++ b/conf.d/python.d/exim.conf
@@ -32,7 +32,13 @@ update_every: 10
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -48,12 +54,13 @@ update_every: 10
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, exim also supports the following:
#
diff --git a/conf.d/python.d/fail2ban.conf b/conf.d/python.d/fail2ban.conf
index 76277108b..60ca87231 100644
--- a/conf.d/python.d/fail2ban.conf
+++ b/conf.d/python.d/fail2ban.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,17 +53,18 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, fail2ban also supports the following:
#
-# log_path: 'path to fail2ban.log' # Default: '/var/log/fail2ban.log'
-# conf_path: 'path to jail.local/jail.conf' # Default: '/etc/fail2ban/jail.local'
-# conf_dir: 'path to jail.d/' # Default: '/etc/fail2ban/jail.d/'
-# exclude: 'jails you want to exclude from autodetection' # Default: '[]' empty list
+# log_path: 'path to fail2ban.log' # Default: '/var/log/fail2ban.log'
+# conf_path: 'path to jail.local/jail.conf' # Default: '/etc/fail2ban/jail.local'
+# conf_dir: 'path to jail.d/' # Default: '/etc/fail2ban/jail.d/'
+# exclude: 'jails you want to exclude from autodetection' # Default: none
#------------------------------------------------------------------------------------------------------------------
diff --git a/conf.d/python.d/freeradius.conf b/conf.d/python.d/freeradius.conf
index b2c8abf6b..3336d4c49 100644
--- a/conf.d/python.d/freeradius.conf
+++ b/conf.d/python.d/freeradius.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,21 +53,22 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, freeradius also supports the following:
#
-# host: 'host' # Default: 'localhost'. Server ip address or hostname.
-# port: 'port' # Default: '18121'. Port on which freeradius server listen (type = status).
+# host: 'host' # Default: 'localhost'. Server ip address or hostname.
+# port: 'port' # Default: '18121'. Port on which freeradius server listen (type = status).
# secret: 'secret' # Default: 'adminsecret'.
-# acct: True/False # Defalt: False. Freeradius accounting statistics.
-# proxy_auth: True/False # Default: False. Freeradius proxy authentication statistics.
-# proxy_acct: True/False # Default: False. Freeradius proxy accounting statistics.
+# acct: yes/no # Default: no. Freeradius accounting statistics.
+# proxy_auth: yes/no # Default: no. Freeradius proxy authentication statistics.
+# proxy_acct: yes/no # Default: no. Freeradius proxy accounting statistics.
#
# ------------------------------------------------------------------------------------------------------------------
# Freeradius server configuration:
@@ -73,14 +80,3 @@
# ln -s ../sites-available/status status
# and restart/reload your FREERADIUS server.
# ------------------------------------------------------------------------------------------------------------------
-#
-# AUTO-DETECTION JOBS
-# only one of them will run (they have the same name)
-
-local:
- host: 'localhost'
- port: '18121'
- secret: 'adminsecret'
-#acct: False
-#proxy_auth: False
-#proxy_acct: False
diff --git a/conf.d/python.d/go_expvar.conf b/conf.d/python.d/go_expvar.conf
index 5be4890dc..c352b1674 100644
--- a/conf.d/python.d/go_expvar.conf
+++ b/conf.d/python.d/go_expvar.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -42,19 +48,17 @@
# predefined parameters. These are:
#
# job_name:
-# name: my_name # the JOB's name as it will appear at the
-# # dashboard. If name: is not supplied the
-# # job_name: will be used (use _ for spaces)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, this plugin also supports the following:
#
# url: 'http://127.0.0.1/debug/vars' # the URL of the expvar endpoint
-# ss_cert: # ignore HTTPS self-signed certificate
-# proxy: # use HTTP proxy
#
# As the plugin cannot possibly know the port your application listens on, there is no default value. Please include
# the whole path of the endpoint, as the expvar handler can be installed in a non-standard location.
diff --git a/conf.d/python.d/haproxy.conf b/conf.d/python.d/haproxy.conf
index a9e048791..a40dd76a5 100644
--- a/conf.d/python.d/haproxy.conf
+++ b/conf.d/python.d/haproxy.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, haproxy also supports the following:
#
diff --git a/conf.d/python.d/hddtemp.conf b/conf.d/python.d/hddtemp.conf
index f74a09803..9165798a2 100644
--- a/conf.d/python.d/hddtemp.conf
+++ b/conf.d/python.d/hddtemp.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,18 +53,18 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, hddtemp also supports the following:
#
# host: 'IP or HOSTNAME' # the host to connect to
# port: PORT # the port to connect to
-# fahrenheit: True/False # fahrenheit instead of celsius. Default is False
#
# By default this module will try to autodetect disks
@@ -78,13 +84,11 @@
localhost:
name: 'local'
host: 'localhost'
- fahrenheit: False
port: 7634
localipv4:
name: 'local'
host: '127.0.0.1'
- fahrenheit: False
port: 7634
localipv6:
diff --git a/conf.d/python.d/ipfs.conf b/conf.d/python.d/ipfs.conf
index e039026cc..c247c1b7a 100644
--- a/conf.d/python.d/ipfs.conf
+++ b/conf.d/python.d/ipfs.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, ipfs also supports the following:
#
diff --git a/conf.d/python.d/isc_dhcpd.conf b/conf.d/python.d/isc_dhcpd.conf
index 938ca6e72..4a4c4a5e3 100644
--- a/conf.d/python.d/isc_dhcpd.conf
+++ b/conf.d/python.d/isc_dhcpd.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, isc_dhcpd supports the following:
#
@@ -71,13 +78,4 @@
# This is the default, so it will work in most cases.
# 3. Pools MUST BE in CIDR format.
#
-#-----------------------------------------------------------------------
-# AUTO-DETECTION JOBS
-# This is disabled by default.
-# To enable it, uncomment the following.
-#
-#leases:
-# leases_path : '/var/lib/dhcp/dhcpd.leases'
-# pools:
-# office: '192.168.2.0/24'
-# wifi: '192.168.3.0/24'
+# ----------------------------------------------------------------------
diff --git a/conf.d/python.d/mdstat.conf b/conf.d/python.d/mdstat.conf
index c89d463be..66a2f153c 100644
--- a/conf.d/python.d/mdstat.conf
+++ b/conf.d/python.d/mdstat.conf
@@ -23,4 +23,10 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
diff --git a/conf.d/python.d/memcached.conf b/conf.d/python.d/memcached.conf
index f1723dc81..85c3daf65 100644
--- a/conf.d/python.d/memcached.conf
+++ b/conf.d/python.d/memcached.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, memcached also supports the following:
#
diff --git a/conf.d/python.d/mongodb.conf b/conf.d/python.d/mongodb.conf
index a19b6570b..62faef68d 100644
--- a/conf.d/python.d/mongodb.conf
+++ b/conf.d/python.d/mongodb.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, mongodb also supports the following:
#
diff --git a/conf.d/python.d/mysql.conf b/conf.d/python.d/mysql.conf
index 63d635174..def9f7e96 100644
--- a/conf.d/python.d/mysql.conf
+++ b/conf.d/python.d/mysql.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, mysql also supports the following:
#
diff --git a/conf.d/python.d/nginx.conf b/conf.d/python.d/nginx.conf
index 645925a55..71c521066 100644
--- a/conf.d/python.d/nginx.conf
+++ b/conf.d/python.d/nginx.conf
@@ -43,7 +43,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -59,13 +65,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: my_name # the JOB's name as it will appear at the
-# # dashboard. If name: is not supplied the
-# # job_name: will be used (use _ for spaces)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, this plugin also supports the following:
#
@@ -99,5 +105,5 @@ localipv4:
localipv6:
name : 'local'
- url : 'http://::1/stub_status'
+ url : 'http://[::1]/stub_status'
diff --git a/conf.d/python.d/nsd.conf b/conf.d/python.d/nsd.conf
index 7566fe85e..078e97216 100644
--- a/conf.d/python.d/nsd.conf
+++ b/conf.d/python.d/nsd.conf
@@ -32,7 +32,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -48,12 +54,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, nsd also supports the following:
#
diff --git a/conf.d/python.d/ovpn_status_log.conf b/conf.d/python.d/ovpn_status_log.conf
index 39bc8e9d4..907f014f5 100644
--- a/conf.d/python.d/ovpn_status_log.conf
+++ b/conf.d/python.d/ovpn_status_log.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, openvpn status log also supports the following:
#
@@ -84,3 +91,5 @@
#
#default:
# log_path: '/var/log/openvpn-status.log'
+#
+# ---------------------------------------------------------------------- \ No newline at end of file
diff --git a/conf.d/python.d/phpfpm.conf b/conf.d/python.d/phpfpm.conf
index f5d067cc7..08688e2fa 100644
--- a/conf.d/python.d/phpfpm.conf
+++ b/conf.d/python.d/phpfpm.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, PHP-FPM also supports the following:
#
diff --git a/conf.d/python.d/postfix.conf b/conf.d/python.d/postfix.conf
index ca9d8fada..e0d5a5f83 100644
--- a/conf.d/python.d/postfix.conf
+++ b/conf.d/python.d/postfix.conf
@@ -32,7 +32,13 @@ update_every: 10
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -48,12 +54,13 @@ update_every: 10
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, postfix also supports the following:
#
diff --git a/conf.d/python.d/postgres.conf b/conf.d/python.d/postgres.conf
index 1dbb64f40..3a70a7184 100644
--- a/conf.d/python.d/postgres.conf
+++ b/conf.d/python.d/postgres.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# A single connection is required in order to pull statistics.
#
@@ -70,7 +77,7 @@
# index_stats : false
# database_poll : 'dbase_name1 dbase_name2' # poll only specified databases (all other will be excluded from charts)
#
-# Postfix permissions are configured at its pg_hba.conf file. You can
+# Postgres permissions are configured at its pg_hba.conf file. You can
# "trust" local clients to allow netdata to connect, or you can create
# a postgres user for netdata and add its password below to allow
# netdata connect.
diff --git a/conf.d/python.d/apache_cache.conf b/conf.d/python.d/powerdns.conf
index 98eecd0e8..ca6200df1 100644
--- a/conf.d/python.d/apache_cache.conf
+++ b/conf.d/python.d/powerdns.conf
@@ -1,4 +1,4 @@
-# netdata python.d.plugin configuration for apache cache
+# netdata python.d.plugin configuration for powerdns
#
# This file is in YaML format. Generally the format is:
#
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,30 +53,26 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
-# Additionally to the above, apache_cache also supports the following:
+# Additionally to the above, apache also supports the following:
#
-# path: 'PATH' # the path to apache's cache.log
+# url: 'URL' # the URL to fetch powerdns performance statistics
+# header:
+# X-API-Key: 'Key' # API key
#
-
# ----------------------------------------------------------------------
# AUTO-DETECTION JOBS
# only one of them will run (they have the same name)
-apache:
- name: 'local'
- path: '/var/log/apache/cache.log'
-
-apache2:
- name: 'local'
- path: '/var/log/apache2/cache.log'
-
-httpd:
- name: 'local'
- path: '/var/log/httpd/cache.log'
+# localhost:
+# name : 'local'
+# url : 'http://127.0.0.1:8081/api/v1/servers/localhost/statistics'
+# header:
+# X-API-Key: 'change_me'
diff --git a/conf.d/python.d/rabbitmq.conf b/conf.d/python.d/rabbitmq.conf
index eccf65df9..3f90da8a2 100644
--- a/conf.d/python.d/rabbitmq.conf
+++ b/conf.d/python.d/rabbitmq.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, rabbitmq plugin also supports the following:
#
diff --git a/conf.d/python.d/redis.conf b/conf.d/python.d/redis.conf
index 983fbfbdb..6363f6da7 100644
--- a/conf.d/python.d/redis.conf
+++ b/conf.d/python.d/redis.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, redis also supports the following:
#
diff --git a/conf.d/python.d/retroshare.conf b/conf.d/python.d/retroshare.conf
index 79614373b..9c92583f7 100644
--- a/conf.d/python.d/retroshare.conf
+++ b/conf.d/python.d/retroshare.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, RetroShare also supports the following:
#
diff --git a/conf.d/python.d/samba.conf b/conf.d/python.d/samba.conf
index 865281cd6..ee513c60f 100644
--- a/conf.d/python.d/samba.conf
+++ b/conf.d/python.d/samba.conf
@@ -31,7 +31,13 @@ update_every: 5
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,10 @@ update_every: 5
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
-#
-#
-
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds \ No newline at end of file
diff --git a/conf.d/python.d/sensors.conf b/conf.d/python.d/sensors.conf
index 2e9a41338..83bbffd7d 100644
--- a/conf.d/python.d/sensors.conf
+++ b/conf.d/python.d/sensors.conf
@@ -23,7 +23,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# Limit the number of sensors types.
@@ -54,5 +60,4 @@ types:
# the prefix is matched (anything that starts like that)
#
#----------------------------------------------------------------------
-# To change celsius to fahrenheit uncomment line below
-#fahrenheit: True
+
diff --git a/conf.d/python.d/smartd_log.conf b/conf.d/python.d/smartd_log.conf
index 8764ffd3e..3fab3f1c0 100644
--- a/conf.d/python.d/smartd_log.conf
+++ b/conf.d/python.d/smartd_log.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,18 +53,20 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, smartd_log also supports the following:
#
-# log_path: '/path/to/smartdlogs' # path to smartd log files. Default is /var/log/smartd
-# raw_values: no # raw or normalized values on charts. Default is normalized.
-# smart_attributes: '1 2 3 4 44' # smart attributes charts. Default are ['1', '4', '5', '7', '9', '12', '193', '194', '197', '198', '200'].
+# log_path: '/path/to/smartdlogs' # path to smartd log files. Default is /var/log/smartd
+# raw_values: yes # enable/disable raw values charts. Enabled by default.
+# smart_attributes: '1 2 3 4 44' # smart attributes charts. Default are ['1', '4', '5', '7', '9', '12', '193', '194', '197', '198', '200'].
+# exclude_disks: 'PATTERN1 PATTERN2' # space separated patterns. If the pattern is in the drive name, the module will not collect data for it.
#
# ----------------------------------------------------------------------
# Additional information
@@ -79,7 +87,4 @@
# Some raw numbers provide valuable insight when properly interpreted. These cases will be discussed later on.
# Raw values are typically listed in hexadecimal numbers. The raw value has different structure for different vendors and is often not meaningful as a decimal number.
#
-#
-# JOB configuration
-#
-log_path: '/var/log/smartd'
+# ----------------------------------------------------------------------
diff --git a/conf.d/python.d/squid.conf b/conf.d/python.d/squid.conf
index 27800bde7..564187f00 100644
--- a/conf.d/python.d/squid.conf
+++ b/conf.d/python.d/squid.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, squid also supports the following:
#
diff --git a/conf.d/python.d/tomcat.conf b/conf.d/python.d/tomcat.conf
index ce89175f6..c63f06cfa 100644
--- a/conf.d/python.d/tomcat.conf
+++ b/conf.d/python.d/tomcat.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,12 +53,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, tomcat also supports the following:
#
diff --git a/conf.d/python.d/varnish.conf b/conf.d/python.d/varnish.conf
index c25f3010f..4b069d514 100644
--- a/conf.d/python.d/varnish.conf
+++ b/conf.d/python.d/varnish.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -47,11 +53,12 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
-#
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
+# ----------------------------------------------------------------------
diff --git a/conf.d/python.d/web_log.conf b/conf.d/python.d/web_log.conf
index cd1f1af00..dd1fff075 100644
--- a/conf.d/python.d/web_log.conf
+++ b/conf.d/python.d/web_log.conf
@@ -31,7 +31,13 @@
# If unset, the default for python.d.plugin is used.
# Attempts to restore the service are made once every update_every
# and only if the module has collected values in the past.
-# retries: 5
+# retries: 60
+
+# autodetection_retry sets the job re-check interval in seconds.
+# The job is not deleted if check fails.
+# Attempts to start the job are made once every autodetection_retry.
+# This feature is disabled by default.
+# autodetection_retry: 0
# ----------------------------------------------------------------------
# JOBS (data collection sources)
@@ -50,12 +56,13 @@
# predefined parameters. These are:
#
# job_name:
-# name: myname # the JOB's name as it will appear at the
-# # dashboard (by default is the job_name)
-# # JOBs sharing a name are mutually exclusive
-# update_every: 1 # the JOB's data collection frequency
-# priority: 60000 # the JOB's order on the dashboard
-# retries: 5 # the JOB's number of restoration attempts
+# name: myname # the JOB's name as it will appear at the
+# # dashboard (by default is the job_name)
+# # JOBs sharing a name are mutually exclusive
+# update_every: 1 # the JOB's data collection frequency
+# priority: 60000 # the JOB's order on the dashboard
+# retries: 60 # the JOB's number of restoration attempts
+# autodetection_retry: 0 # the JOB's re-check interval in seconds
#
# Additionally to the above, web_log also supports the following:
#
diff --git a/conf.d/stream.conf b/conf.d/stream.conf
index 0ae5ba673..8945529ee 100644
--- a/conf.d/stream.conf
+++ b/conf.d/stream.conf
@@ -21,13 +21,13 @@
#
# If many are given, the first available will get the metrics.
#
- # PROTOCOL = tcp or udp (only tcp is supported by masters)
- # HOST = an IPv4, IPv6 IP, or a hostname.
+ # PROTOCOL = tcp, udp, or unix (only tcp and unix are supported by masters)
+ # HOST = an IPv4, IPv6 IP, or a hostname, or a unix domain socket path.
# IPv6 IPs should be given with brackets [ip:address]
- # INTERFACE = the network interface to use
+ # INTERFACE = the network interface to use (only for IPv6)
# PORT = the port number or service name (/etc/services)
#
- # This communication is not HTTP (cannot be proxied by web proxies).
+ # This communication is not HTTP (it cannot be proxied by web proxies).
destination =
# The API_KEY to use (as the sender)
@@ -56,26 +56,34 @@
# -----------------------------------------------------------------------------
# 2. ON MASTER NETDATA - THE ONE THAT WILL BE RECEIVING METRICS
-# You can have one API key per slave, or the same API key for all slaves.
+# You can have one API key per slave,
+# or the same API key for all slaves.
#
# netdata searches for options in this order:
#
-# a) [MACHINE_GUID] section (settings for each machine)
-# b) [API_KEY] section (settings for the API key)
-# c) master netdata settings (netdata.conf)
+# a) master netdata settings (netdata.conf)
+# b) [API_KEY] section (below, settings for the API key)
+# c) [MACHINE_GUID] section (below, settings for each machine)
#
# You can combine the above (the more specific setting will be used).
# API key authentication
-# If the key is not listed here, it will not be able to connect.
+# If the key is not listed here, it will not be able to push metrics.
+# [API_KEY] is [YOUR-API-KEY], i.e [11111111-2222-3333-4444-555555555555]
[API_KEY]
- # Default settings for the API key
+ # Default settings for this API key
# You can disable the API key, by setting this to: no
# The default (for unknown API keys) is: no
enabled = no
+ # A list of simple patterns matching the IPs of the servers that
+ # will be pushing metrics using this API key.
+ # The metrics are received via the API port, so the same IPs
+ # should also be matched at netdata.conf [web].allow connections from
+ allow from = *
+
# The default history in entries, for all hosts using this API key.
# You can also set it per host below.
# If you don't set it here, the history size of the central netdata
@@ -132,6 +140,13 @@
# Use only the API key for security.
enabled = no
+ # A list of simple patterns matching the IPs of the servers that
+ # will be pushing metrics using this MACHINE GUID.
+ # The metrics are received via the API port, so the same IPs
+ # should also be matched at netdata.conf [web].allow connections from
+ # and at stream.conf [API_KEY].allow from
+ allow from = *
+
# The number of entries in the database
history = 3600