summaryrefslogtreecommitdiffstats
path: root/collectors
diff options
context:
space:
mode:
Diffstat (limited to '')
-rw-r--r--collectors/COLLECTORS.md1194
-rw-r--r--collectors/Makefile.am42
-rw-r--r--collectors/README.md62
-rw-r--r--collectors/REFERENCE.md149
-rw-r--r--collectors/apps.plugin/Makefile.am12
-rw-r--r--collectors/apps.plugin/README.md402
-rw-r--r--collectors/apps.plugin/apps_plugin.c5422
-rw-r--r--collectors/cgroups.plugin/Makefile.am23
-rw-r--r--collectors/cgroups.plugin/README.md302
-rw-r--r--collectors/cgroups.plugin/metadata.yaml1022
-rw-r--r--collectors/charts.d.plugin/Makefile.am49
-rw-r--r--collectors/charts.d.plugin/README.md190
-rw-r--r--collectors/charts.d.plugin/ap/Makefile.inc13
-rw-r--r--collectors/charts.d.plugin/apcupsd/Makefile.inc13
-rw-r--r--collectors/charts.d.plugin/apcupsd/metadata.yaml256
-rw-r--r--collectors/charts.d.plugin/example/Makefile.inc13
-rw-r--r--collectors/charts.d.plugin/example/README.md14
-rw-r--r--collectors/charts.d.plugin/libreswan/Makefile.inc13
-rw-r--r--collectors/charts.d.plugin/opensips/Makefile.inc13
-rw-r--r--collectors/charts.d.plugin/sensors/Makefile.inc13
-rw-r--r--collectors/charts.d.plugin/sensors/metadata.yaml182
-rw-r--r--collectors/checks.plugin/README.md12
-rw-r--r--collectors/cups.plugin/Makefile.am8
-rw-r--r--collectors/debugfs.plugin/Makefile.am9
-rw-r--r--collectors/debugfs.plugin/README.md65
-rw-r--r--collectors/diskspace.plugin/Makefile.am8
-rw-r--r--collectors/diskspace.plugin/metadata.yaml139
-rw-r--r--collectors/ebpf.plugin/Makefile.am42
-rw-r--r--collectors/ebpf.plugin/README.md1071
-rw-r--r--collectors/ebpf.plugin/ebpf.c4126
-rw-r--r--collectors/ebpf.plugin/ebpf.d/swap.conf34
-rw-r--r--collectors/ebpf.plugin/ebpf.h393
-rw-r--r--collectors/ebpf.plugin/ebpf_cgroup.h71
-rw-r--r--collectors/ebpf.plugin/ebpf_functions.c1093
-rw-r--r--collectors/ebpf.plugin/ebpf_functions.h44
-rw-r--r--collectors/ebpf.plugin/metadata.yaml3320
-rw-r--r--collectors/freebsd.plugin/Makefile.am8
-rw-r--r--collectors/freebsd.plugin/README.md16
-rw-r--r--collectors/freebsd.plugin/metadata.yaml3398
-rw-r--r--collectors/freeipmi.plugin/Makefile.am8
-rw-r--r--collectors/freeipmi.plugin/metadata.yaml347
-rw-r--r--collectors/idlejitter.plugin/Makefile.am8
-rw-r--r--collectors/ioping.plugin/Makefile.am24
-rw-r--r--collectors/ioping.plugin/metadata.yaml101
-rw-r--r--collectors/log2journal/Makefile.am17
-rw-r--r--collectors/log2journal/README.md912
-rw-r--r--collectors/macos.plugin/Makefile.am8
-rw-r--r--collectors/macos.plugin/metadata.yaml727
-rw-r--r--collectors/nfacct.plugin/Makefile.am8
-rw-r--r--collectors/perf.plugin/Makefile.am8
-rw-r--r--collectors/perf.plugin/metadata.yaml252
-rw-r--r--collectors/plugins.d/Makefile.am12
-rw-r--r--collectors/plugins.d/README.md680
-rw-r--r--collectors/plugins.d/gperf-config.txt58
-rw-r--r--collectors/plugins.d/gperf-hashtable.h177
-rw-r--r--collectors/plugins.d/local_listeners.c400
-rw-r--r--collectors/plugins.d/plugins_d.h67
-rw-r--r--collectors/plugins.d/pluginsd_parser.c3208
-rw-r--r--collectors/proc.plugin/Makefile.am8
-rw-r--r--collectors/proc.plugin/README.md639
-rw-r--r--collectors/proc.plugin/integrations/btrfs.md137
-rw-r--r--collectors/proc.plugin/integrations/network_interfaces.md137
-rw-r--r--collectors/proc.plugin/metadata.yaml5299
-rw-r--r--collectors/profile.plugin/Makefile.am8
-rw-r--r--collectors/python.d.plugin/Makefile.am229
-rw-r--r--collectors/python.d.plugin/README.md77
-rw-r--r--collectors/python.d.plugin/adaptec_raid/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/adaptec_raid/metadata.yaml167
-rw-r--r--collectors/python.d.plugin/alarms/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/alarms/metadata.yaml177
-rw-r--r--collectors/python.d.plugin/am2320/Makefile.inc8
-rw-r--r--collectors/python.d.plugin/anomalies/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/anomalies/README.md248
-rw-r--r--collectors/python.d.plugin/anomalies/metadata.yaml87
-rw-r--r--collectors/python.d.plugin/beanstalk/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/beanstalk/metadata.yaml263
-rw-r--r--collectors/python.d.plugin/bind_rndc/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/bind_rndc/metadata.yaml191
-rw-r--r--collectors/python.d.plugin/boinc/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/boinc/metadata.yaml198
-rw-r--r--collectors/python.d.plugin/ceph/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/ceph/metadata.yaml223
-rw-r--r--collectors/python.d.plugin/changefinder/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/dovecot/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/example/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/example/metadata.yaml138
-rw-r--r--collectors/python.d.plugin/exim/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/fail2ban/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/gearman/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/gearman/metadata.yaml168
-rw-r--r--collectors/python.d.plugin/go_expvar/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/go_expvar/metadata.yaml329
-rw-r--r--collectors/python.d.plugin/haproxy/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/haproxy/README.md90
-rw-r--r--collectors/python.d.plugin/haproxy/metadata.yaml322
-rw-r--r--collectors/python.d.plugin/hddtemp/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/hpssa/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/hpssa/metadata.yaml185
-rw-r--r--collectors/python.d.plugin/icecast/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/ipfs/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/ipfs/metadata.yaml172
-rw-r--r--collectors/python.d.plugin/litespeed/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/megacli/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/megacli/metadata.yaml193
-rw-r--r--collectors/python.d.plugin/memcached/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/memcached/metadata.yaml247
-rw-r--r--collectors/python.d.plugin/monit/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/nsd/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/nvidia_smi/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/nvidia_smi/README.md157
-rw-r--r--collectors/python.d.plugin/nvidia_smi/metadata.yaml166
-rw-r--r--collectors/python.d.plugin/openldap/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/oracledb/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/pandas/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/postfix/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/puppet/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/rethinkdbs/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/retroshare/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/retroshare/metadata.yaml144
-rw-r--r--collectors/python.d.plugin/riakkv/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/riakkv/metadata.yaml358
-rw-r--r--collectors/python.d.plugin/samba/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/samba/metadata.yaml205
-rw-r--r--collectors/python.d.plugin/sensors/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/sensors/metadata.yaml184
-rw-r--r--collectors/python.d.plugin/smartd_log/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/spigotmc/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/squid/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/tomcat/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/tor/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/traefik/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/traefik/README.md98
-rw-r--r--collectors/python.d.plugin/traefik/metadata.yaml125
-rw-r--r--collectors/python.d.plugin/uwsgi/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/varnish/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/w1sensor/Makefile.inc13
-rw-r--r--collectors/python.d.plugin/zscores/Makefile.inc12
-rw-r--r--collectors/python.d.plugin/zscores/metadata.yaml187
-rw-r--r--collectors/slabinfo.plugin/Makefile.am14
-rw-r--r--collectors/statsd.plugin/Makefile.am23
-rw-r--r--collectors/statsd.plugin/README.md1052
-rw-r--r--collectors/systemd-journal.plugin/Makefile.am16
-rw-r--r--collectors/systemd-journal.plugin/README.md472
-rw-r--r--collectors/systemd-journal.plugin/systemd-journal-annotations.c719
-rw-r--r--collectors/systemd-journal.plugin/systemd-main.c112
-rw-r--r--collectors/tc.plugin/Makefile.am20
-rw-r--r--collectors/timex.plugin/Makefile.am8
-rw-r--r--collectors/timex.plugin/metadata.yaml112
-rw-r--r--collectors/xenstat.plugin/Makefile.am8
-rw-r--r--docs/observability-centralization-points/logs-centralization-points-with-systemd-journald/active-journal-source-without-encryption.md (renamed from collectors/systemd-journal.plugin/active_journal_centralization_guide_no_encryption.md)0
-rw-r--r--docs/observability-centralization-points/logs-centralization-points-with-systemd-journald/passive-journal-centralization-without-encryption.md (renamed from collectors/systemd-journal.plugin/passive_journal_centralization_guide_no_encryption.md)0
-rw-r--r--src/collectors/all.h (renamed from collectors/all.h)18
-rw-r--r--src/collectors/apps.plugin/apps_groups.conf (renamed from collectors/apps.plugin/apps_groups.conf)1
-rw-r--r--src/collectors/apps.plugin/integrations/applications.md (renamed from collectors/apps.plugin/integrations/applications.md)6
-rw-r--r--src/collectors/apps.plugin/integrations/user_groups.md (renamed from collectors/apps.plugin/integrations/user_groups.md)6
-rw-r--r--src/collectors/apps.plugin/integrations/users.md (renamed from collectors/apps.plugin/integrations/users.md)6
-rw-r--r--src/collectors/apps.plugin/metadata.yaml (renamed from collectors/apps.plugin/metadata.yaml)0
-rw-r--r--src/collectors/cgroups.plugin/cgroup-charts.c (renamed from collectors/cgroups.plugin/cgroup-charts.c)2
-rw-r--r--src/collectors/cgroups.plugin/cgroup-discovery.c (renamed from collectors/cgroups.plugin/cgroup-discovery.c)94
-rw-r--r--src/collectors/cgroups.plugin/cgroup-internals.h (renamed from collectors/cgroups.plugin/cgroup-internals.h)21
-rwxr-xr-xsrc/collectors/cgroups.plugin/cgroup-name.sh.in (renamed from collectors/cgroups.plugin/cgroup-name.sh.in)0
-rwxr-xr-xsrc/collectors/cgroups.plugin/cgroup-network-helper.sh.in (renamed from collectors/cgroups.plugin/cgroup-network-helper.sh.in)0
-rw-r--r--src/collectors/cgroups.plugin/cgroup-network.c (renamed from collectors/cgroups.plugin/cgroup-network.c)3
-rw-r--r--src/collectors/cgroups.plugin/cgroup-top.c (renamed from collectors/cgroups.plugin/cgroup-top.c)52
-rw-r--r--src/collectors/cgroups.plugin/integrations/containers.md (renamed from collectors/cgroups.plugin/integrations/containers.md)14
-rw-r--r--src/collectors/cgroups.plugin/integrations/kubernetes_containers.md (renamed from collectors/cgroups.plugin/integrations/kubernetes_containers.md)14
-rw-r--r--src/collectors/cgroups.plugin/integrations/libvirt_containers.md (renamed from collectors/cgroups.plugin/integrations/libvirt_containers.md)14
-rw-r--r--src/collectors/cgroups.plugin/integrations/lxc_containers.md (renamed from collectors/cgroups.plugin/integrations/lxc_containers.md)14
-rw-r--r--src/collectors/cgroups.plugin/integrations/ovirt_containers.md (renamed from collectors/cgroups.plugin/integrations/ovirt_containers.md)14
-rw-r--r--src/collectors/cgroups.plugin/integrations/proxmox_containers.md (renamed from collectors/cgroups.plugin/integrations/proxmox_containers.md)14
-rw-r--r--src/collectors/cgroups.plugin/integrations/systemd_services.md (renamed from collectors/cgroups.plugin/integrations/systemd_services.md)6
-rw-r--r--src/collectors/cgroups.plugin/integrations/virtual_machines.md (renamed from collectors/cgroups.plugin/integrations/virtual_machines.md)14
-rw-r--r--src/collectors/cgroups.plugin/sys_fs_cgroup.c (renamed from collectors/cgroups.plugin/sys_fs_cgroup.c)116
-rw-r--r--src/collectors/cgroups.plugin/sys_fs_cgroup.h (renamed from collectors/cgroups.plugin/sys_fs_cgroup.h)0
-rw-r--r--src/collectors/cgroups.plugin/tests/test_cgroups_plugin.c (renamed from collectors/cgroups.plugin/tests/test_cgroups_plugin.c)0
-rw-r--r--src/collectors/cgroups.plugin/tests/test_cgroups_plugin.h (renamed from collectors/cgroups.plugin/tests/test_cgroups_plugin.h)0
-rw-r--r--src/collectors/cgroups.plugin/tests/test_doubles.c (renamed from collectors/cgroups.plugin/tests/test_doubles.c)4
l---------src/collectors/charts.d.plugin/ap/README.md (renamed from collectors/charts.d.plugin/ap/README.md)0
-rw-r--r--src/collectors/charts.d.plugin/ap/ap.chart.sh (renamed from collectors/charts.d.plugin/ap/ap.chart.sh)0
-rw-r--r--src/collectors/charts.d.plugin/ap/ap.conf (renamed from collectors/charts.d.plugin/ap/ap.conf)0
-rw-r--r--src/collectors/charts.d.plugin/ap/integrations/access_points.md (renamed from collectors/charts.d.plugin/ap/integrations/access_points.md)8
-rw-r--r--src/collectors/charts.d.plugin/ap/metadata.yaml (renamed from collectors/charts.d.plugin/ap/metadata.yaml)0
l---------src/collectors/charts.d.plugin/apcupsd/README.md (renamed from collectors/charts.d.plugin/apcupsd/README.md)0
-rw-r--r--src/collectors/charts.d.plugin/apcupsd/apcupsd.chart.sh (renamed from collectors/charts.d.plugin/apcupsd/apcupsd.chart.sh)0
-rw-r--r--src/collectors/charts.d.plugin/apcupsd/apcupsd.conf (renamed from collectors/charts.d.plugin/apcupsd/apcupsd.conf)0
-rw-r--r--src/collectors/charts.d.plugin/apcupsd/integrations/apc_ups.md (renamed from collectors/charts.d.plugin/apcupsd/integrations/apc_ups.md)28
-rw-r--r--src/collectors/charts.d.plugin/charts.d.conf (renamed from collectors/charts.d.plugin/charts.d.conf)0
-rwxr-xr-xsrc/collectors/charts.d.plugin/charts.d.dryrun-helper.sh (renamed from collectors/charts.d.plugin/charts.d.dryrun-helper.sh)0
-rwxr-xr-xsrc/collectors/charts.d.plugin/charts.d.plugin.in (renamed from collectors/charts.d.plugin/charts.d.plugin.in)0
-rw-r--r--src/collectors/charts.d.plugin/example/example.chart.sh (renamed from collectors/charts.d.plugin/example/example.chart.sh)0
-rw-r--r--src/collectors/charts.d.plugin/example/example.conf (renamed from collectors/charts.d.plugin/example/example.conf)0
l---------src/collectors/charts.d.plugin/libreswan/README.md (renamed from collectors/charts.d.plugin/libreswan/README.md)0
-rw-r--r--src/collectors/charts.d.plugin/libreswan/integrations/libreswan.md (renamed from collectors/charts.d.plugin/libreswan/integrations/libreswan.md)8
-rw-r--r--src/collectors/charts.d.plugin/libreswan/libreswan.chart.sh (renamed from collectors/charts.d.plugin/libreswan/libreswan.chart.sh)0
-rw-r--r--src/collectors/charts.d.plugin/libreswan/libreswan.conf (renamed from collectors/charts.d.plugin/libreswan/libreswan.conf)0
-rw-r--r--src/collectors/charts.d.plugin/libreswan/metadata.yaml (renamed from collectors/charts.d.plugin/libreswan/metadata.yaml)0
-rw-r--r--src/collectors/charts.d.plugin/loopsleepms.sh.inc (renamed from collectors/charts.d.plugin/loopsleepms.sh.inc)0
l---------src/collectors/charts.d.plugin/opensips/README.md (renamed from collectors/charts.d.plugin/opensips/README.md)0
-rw-r--r--src/collectors/charts.d.plugin/opensips/integrations/opensips.md (renamed from collectors/charts.d.plugin/opensips/integrations/opensips.md)8
-rw-r--r--src/collectors/charts.d.plugin/opensips/metadata.yaml (renamed from collectors/charts.d.plugin/opensips/metadata.yaml)0
-rw-r--r--src/collectors/charts.d.plugin/opensips/opensips.chart.sh (renamed from collectors/charts.d.plugin/opensips/opensips.chart.sh)0
-rw-r--r--src/collectors/charts.d.plugin/opensips/opensips.conf (renamed from collectors/charts.d.plugin/opensips/opensips.conf)0
l---------src/collectors/charts.d.plugin/sensors/README.md (renamed from collectors/charts.d.plugin/sensors/README.md)0
-rw-r--r--src/collectors/charts.d.plugin/sensors/integrations/linux_sensors_sysfs.md (renamed from collectors/charts.d.plugin/sensors/integrations/linux_sensors_sysfs.md)14
-rw-r--r--src/collectors/charts.d.plugin/sensors/sensors.chart.sh (renamed from collectors/charts.d.plugin/sensors/sensors.chart.sh)0
-rw-r--r--src/collectors/charts.d.plugin/sensors/sensors.conf (renamed from collectors/charts.d.plugin/sensors/sensors.conf)0
l---------src/collectors/cups.plugin/README.md (renamed from collectors/cups.plugin/README.md)0
-rw-r--r--src/collectors/cups.plugin/cups_plugin.c (renamed from collectors/cups.plugin/cups_plugin.c)2
-rw-r--r--src/collectors/cups.plugin/integrations/cups.md (renamed from collectors/cups.plugin/integrations/cups.md)8
-rw-r--r--src/collectors/cups.plugin/metadata.yaml (renamed from collectors/cups.plugin/metadata.yaml)0
-rw-r--r--src/collectors/debugfs.plugin/debugfs_extfrag.c (renamed from collectors/debugfs.plugin/debugfs_extfrag.c)0
-rw-r--r--src/collectors/debugfs.plugin/debugfs_plugin.c (renamed from collectors/debugfs.plugin/debugfs_plugin.c)4
-rw-r--r--src/collectors/debugfs.plugin/debugfs_plugin.h (renamed from collectors/debugfs.plugin/debugfs_plugin.h)0
-rw-r--r--src/collectors/debugfs.plugin/debugfs_zswap.c (renamed from collectors/debugfs.plugin/debugfs_zswap.c)2
-rw-r--r--src/collectors/debugfs.plugin/integrations/linux_zswap.md (renamed from collectors/debugfs.plugin/integrations/linux_zswap.md)8
-rw-r--r--src/collectors/debugfs.plugin/integrations/power_capping.md (renamed from collectors/debugfs.plugin/integrations/power_capping.md)8
-rw-r--r--src/collectors/debugfs.plugin/integrations/system_memory_fragmentation.md (renamed from collectors/debugfs.plugin/integrations/system_memory_fragmentation.md)8
-rw-r--r--src/collectors/debugfs.plugin/metadata.yaml (renamed from collectors/debugfs.plugin/metadata.yaml)0
-rw-r--r--src/collectors/debugfs.plugin/sys_devices_virtual_powercap.c (renamed from collectors/debugfs.plugin/sys_devices_virtual_powercap.c)2
l---------src/collectors/diskspace.plugin/README.md (renamed from collectors/diskspace.plugin/README.md)0
-rw-r--r--src/collectors/diskspace.plugin/integrations/disk_space.md (renamed from collectors/diskspace.plugin/integrations/disk_space.md)12
-rw-r--r--src/collectors/diskspace.plugin/plugin_diskspace.c (renamed from collectors/diskspace.plugin/plugin_diskspace.c)212
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d.conf (renamed from collectors/ebpf.plugin/ebpf.d.conf)10
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/cachestat.conf (renamed from collectors/ebpf.plugin/ebpf.d/cachestat.conf)2
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/dcstat.conf (renamed from collectors/ebpf.plugin/ebpf.d/dcstat.conf)2
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/disk.conf (renamed from collectors/ebpf.plugin/ebpf.d/disk.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/ebpf_kernel_reject_list.txt (renamed from collectors/ebpf.plugin/ebpf.d/ebpf_kernel_reject_list.txt)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/fd.conf (renamed from collectors/ebpf.plugin/ebpf.d/fd.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/filesystem.conf (renamed from collectors/ebpf.plugin/ebpf.d/filesystem.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/functions.conf (renamed from collectors/ebpf.plugin/ebpf.d/functions.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/hardirq.conf (renamed from collectors/ebpf.plugin/ebpf.d/hardirq.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/mdflush.conf (renamed from collectors/ebpf.plugin/ebpf.d/mdflush.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/mount.conf (renamed from collectors/ebpf.plugin/ebpf.d/mount.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/network.conf (renamed from collectors/ebpf.plugin/ebpf.d/network.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/oomkill.conf (renamed from collectors/ebpf.plugin/ebpf.d/oomkill.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/process.conf (renamed from collectors/ebpf.plugin/ebpf.d/process.conf)2
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/shm.conf (renamed from collectors/ebpf.plugin/ebpf.d/shm.conf)1
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/softirq.conf (renamed from collectors/ebpf.plugin/ebpf.d/softirq.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/sync.conf (renamed from collectors/ebpf.plugin/ebpf.d/sync.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf.d/vfs.conf (renamed from collectors/ebpf.plugin/ebpf.d/vfs.conf)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_apps.c (renamed from collectors/ebpf.plugin/ebpf_apps.c)445
-rw-r--r--src/collectors/ebpf.plugin/ebpf_apps.h (renamed from collectors/ebpf.plugin/ebpf_apps.h)110
-rw-r--r--src/collectors/ebpf.plugin/ebpf_cachestat.c (renamed from collectors/ebpf.plugin/ebpf_cachestat.c)481
-rw-r--r--src/collectors/ebpf.plugin/ebpf_cachestat.h (renamed from collectors/ebpf.plugin/ebpf_cachestat.h)20
-rw-r--r--src/collectors/ebpf.plugin/ebpf_cgroup.c (renamed from collectors/ebpf.plugin/ebpf_cgroup.c)25
-rw-r--r--src/collectors/ebpf.plugin/ebpf_dcstat.c (renamed from collectors/ebpf.plugin/ebpf_dcstat.c)585
-rw-r--r--src/collectors/ebpf.plugin/ebpf_dcstat.h (renamed from collectors/ebpf.plugin/ebpf_dcstat.h)16
-rw-r--r--src/collectors/ebpf.plugin/ebpf_disk.c (renamed from collectors/ebpf.plugin/ebpf_disk.c)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_disk.h (renamed from collectors/ebpf.plugin/ebpf_disk.h)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_fd.c (renamed from collectors/ebpf.plugin/ebpf_fd.c)466
-rw-r--r--src/collectors/ebpf.plugin/ebpf_fd.h (renamed from collectors/ebpf.plugin/ebpf_fd.h)14
-rw-r--r--src/collectors/ebpf.plugin/ebpf_filesystem.c (renamed from collectors/ebpf.plugin/ebpf_filesystem.c)2
-rw-r--r--src/collectors/ebpf.plugin/ebpf_filesystem.h (renamed from collectors/ebpf.plugin/ebpf_filesystem.h)3
-rw-r--r--src/collectors/ebpf.plugin/ebpf_hardirq.c (renamed from collectors/ebpf.plugin/ebpf_hardirq.c)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_hardirq.h (renamed from collectors/ebpf.plugin/ebpf_hardirq.h)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_mdflush.c (renamed from collectors/ebpf.plugin/ebpf_mdflush.c)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_mdflush.h (renamed from collectors/ebpf.plugin/ebpf_mdflush.h)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_mount.c (renamed from collectors/ebpf.plugin/ebpf_mount.c)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_mount.h (renamed from collectors/ebpf.plugin/ebpf_mount.h)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_oomkill.c (renamed from collectors/ebpf.plugin/ebpf_oomkill.c)71
-rw-r--r--src/collectors/ebpf.plugin/ebpf_oomkill.h (renamed from collectors/ebpf.plugin/ebpf_oomkill.h)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_process.c (renamed from collectors/ebpf.plugin/ebpf_process.c)335
-rw-r--r--src/collectors/ebpf.plugin/ebpf_process.h (renamed from collectors/ebpf.plugin/ebpf_process.h)11
-rw-r--r--src/collectors/ebpf.plugin/ebpf_shm.c (renamed from collectors/ebpf.plugin/ebpf_shm.c)403
-rw-r--r--src/collectors/ebpf.plugin/ebpf_shm.h (renamed from collectors/ebpf.plugin/ebpf_shm.h)14
-rw-r--r--src/collectors/ebpf.plugin/ebpf_socket.c (renamed from collectors/ebpf.plugin/ebpf_socket.c)515
-rw-r--r--src/collectors/ebpf.plugin/ebpf_socket.h (renamed from collectors/ebpf.plugin/ebpf_socket.h)24
-rw-r--r--src/collectors/ebpf.plugin/ebpf_softirq.c (renamed from collectors/ebpf.plugin/ebpf_softirq.c)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_softirq.h (renamed from collectors/ebpf.plugin/ebpf_softirq.h)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_swap.c (renamed from collectors/ebpf.plugin/ebpf_swap.c)365
-rw-r--r--src/collectors/ebpf.plugin/ebpf_swap.h (renamed from collectors/ebpf.plugin/ebpf_swap.h)14
-rw-r--r--src/collectors/ebpf.plugin/ebpf_sync.c (renamed from collectors/ebpf.plugin/ebpf_sync.c)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_sync.h (renamed from collectors/ebpf.plugin/ebpf_sync.h)4
-rw-r--r--src/collectors/ebpf.plugin/ebpf_unittest.c (renamed from collectors/ebpf.plugin/ebpf_unittest.c)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_unittest.h (renamed from collectors/ebpf.plugin/ebpf_unittest.h)0
-rw-r--r--src/collectors/ebpf.plugin/ebpf_vfs.c (renamed from collectors/ebpf.plugin/ebpf_vfs.c)765
-rw-r--r--src/collectors/ebpf.plugin/ebpf_vfs.h (renamed from collectors/ebpf.plugin/ebpf_vfs.h)35
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_cachestat.md (renamed from collectors/ebpf.plugin/integrations/ebpf_cachestat.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_dcstat.md (renamed from collectors/ebpf.plugin/integrations/ebpf_dcstat.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_disk.md (renamed from collectors/ebpf.plugin/integrations/ebpf_disk.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_filedescriptor.md (renamed from collectors/ebpf.plugin/integrations/ebpf_filedescriptor.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_filesystem.md (renamed from collectors/ebpf.plugin/integrations/ebpf_filesystem.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_hardirq.md (renamed from collectors/ebpf.plugin/integrations/ebpf_hardirq.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_mdflush.md (renamed from collectors/ebpf.plugin/integrations/ebpf_mdflush.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_mount.md (renamed from collectors/ebpf.plugin/integrations/ebpf_mount.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_oomkill.md (renamed from collectors/ebpf.plugin/integrations/ebpf_oomkill.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_process.md (renamed from collectors/ebpf.plugin/integrations/ebpf_process.md)6
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_processes.md (renamed from collectors/ebpf.plugin/integrations/ebpf_processes.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_shm.md (renamed from collectors/ebpf.plugin/integrations/ebpf_shm.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_socket.md (renamed from collectors/ebpf.plugin/integrations/ebpf_socket.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_softirq.md (renamed from collectors/ebpf.plugin/integrations/ebpf_softirq.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_swap.md (renamed from collectors/ebpf.plugin/integrations/ebpf_swap.md)8
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_sync.md (renamed from collectors/ebpf.plugin/integrations/ebpf_sync.md)10
-rw-r--r--src/collectors/ebpf.plugin/integrations/ebpf_vfs.md (renamed from collectors/ebpf.plugin/integrations/ebpf_vfs.md)8
-rw-r--r--src/collectors/freebsd.plugin/freebsd_devstat.c (renamed from collectors/freebsd.plugin/freebsd_devstat.c)0
-rw-r--r--src/collectors/freebsd.plugin/freebsd_getifaddrs.c (renamed from collectors/freebsd.plugin/freebsd_getifaddrs.c)0
-rw-r--r--src/collectors/freebsd.plugin/freebsd_getmntinfo.c (renamed from collectors/freebsd.plugin/freebsd_getmntinfo.c)0
-rw-r--r--src/collectors/freebsd.plugin/freebsd_ipfw.c (renamed from collectors/freebsd.plugin/freebsd_ipfw.c)0
-rw-r--r--src/collectors/freebsd.plugin/freebsd_kstat_zfs.c (renamed from collectors/freebsd.plugin/freebsd_kstat_zfs.c)0
-rw-r--r--src/collectors/freebsd.plugin/freebsd_sysctl.c (renamed from collectors/freebsd.plugin/freebsd_sysctl.c)0
-rw-r--r--src/collectors/freebsd.plugin/integrations/dev.cpu.0.freq.md (renamed from collectors/freebsd.plugin/integrations/dev.cpu.0.freq.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/dev.cpu.temperature.md (renamed from collectors/freebsd.plugin/integrations/dev.cpu.temperature.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/devstat.md (renamed from collectors/freebsd.plugin/integrations/devstat.md)10
-rw-r--r--src/collectors/freebsd.plugin/integrations/getifaddrs.md (renamed from collectors/freebsd.plugin/integrations/getifaddrs.md)22
-rw-r--r--src/collectors/freebsd.plugin/integrations/getmntinfo.md (renamed from collectors/freebsd.plugin/integrations/getmntinfo.md)12
-rw-r--r--src/collectors/freebsd.plugin/integrations/hw.intrcnt.md (renamed from collectors/freebsd.plugin/integrations/hw.intrcnt.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/ipfw.md (renamed from collectors/freebsd.plugin/integrations/ipfw.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/kern.cp_time.md (renamed from collectors/freebsd.plugin/integrations/kern.cp_time.md)16
-rw-r--r--src/collectors/freebsd.plugin/integrations/kern.ipc.msq.md (renamed from collectors/freebsd.plugin/integrations/kern.ipc.msq.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/kern.ipc.sem.md (renamed from collectors/freebsd.plugin/integrations/kern.ipc.sem.md)12
-rw-r--r--src/collectors/freebsd.plugin/integrations/kern.ipc.shm.md (renamed from collectors/freebsd.plugin/integrations/kern.ipc.shm.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/net.inet.icmp.stats.md (renamed from collectors/freebsd.plugin/integrations/net.inet.icmp.stats.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/net.inet.ip.stats.md (renamed from collectors/freebsd.plugin/integrations/net.inet.ip.stats.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/net.inet.tcp.states.md (renamed from collectors/freebsd.plugin/integrations/net.inet.tcp.states.md)10
-rw-r--r--src/collectors/freebsd.plugin/integrations/net.inet.tcp.stats.md (renamed from collectors/freebsd.plugin/integrations/net.inet.tcp.stats.md)16
-rw-r--r--src/collectors/freebsd.plugin/integrations/net.inet.udp.stats.md (renamed from collectors/freebsd.plugin/integrations/net.inet.udp.stats.md)12
-rw-r--r--src/collectors/freebsd.plugin/integrations/net.inet6.icmp6.stats.md (renamed from collectors/freebsd.plugin/integrations/net.inet6.icmp6.stats.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/net.inet6.ip6.stats.md (renamed from collectors/freebsd.plugin/integrations/net.inet6.ip6.stats.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/net.isr.md (renamed from collectors/freebsd.plugin/integrations/net.isr.md)14
-rw-r--r--src/collectors/freebsd.plugin/integrations/system.ram.md (renamed from collectors/freebsd.plugin/integrations/system.ram.md)16
-rw-r--r--src/collectors/freebsd.plugin/integrations/uptime.md (renamed from collectors/freebsd.plugin/integrations/uptime.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/vm.loadavg.md (renamed from collectors/freebsd.plugin/integrations/vm.loadavg.md)16
-rw-r--r--src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_intr.md (renamed from collectors/freebsd.plugin/integrations/vm.stats.sys.v_intr.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_soft.md (renamed from collectors/freebsd.plugin/integrations/vm.stats.sys.v_soft.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_swtch.md (renamed from collectors/freebsd.plugin/integrations/vm.stats.sys.v_swtch.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/vm.stats.vm.v_pgfaults.md (renamed from collectors/freebsd.plugin/integrations/vm.stats.vm.v_pgfaults.md)8
-rw-r--r--src/collectors/freebsd.plugin/integrations/vm.stats.vm.v_swappgs.md (renamed from collectors/freebsd.plugin/integrations/vm.stats.vm.v_swappgs.md)10
-rw-r--r--src/collectors/freebsd.plugin/integrations/vm.swap_info.md (renamed from collectors/freebsd.plugin/integrations/vm.swap_info.md)10
-rw-r--r--src/collectors/freebsd.plugin/integrations/vm.vmtotal.md (renamed from collectors/freebsd.plugin/integrations/vm.vmtotal.md)10
-rw-r--r--src/collectors/freebsd.plugin/integrations/zfs.md (renamed from collectors/freebsd.plugin/integrations/zfs.md)10
-rw-r--r--src/collectors/freebsd.plugin/plugin_freebsd.c (renamed from collectors/freebsd.plugin/plugin_freebsd.c)2
-rw-r--r--src/collectors/freebsd.plugin/plugin_freebsd.h (renamed from collectors/freebsd.plugin/plugin_freebsd.h)0
l---------src/collectors/freeipmi.plugin/README.md (renamed from collectors/freeipmi.plugin/README.md)0
-rw-r--r--src/collectors/freeipmi.plugin/freeipmi_plugin.c (renamed from collectors/freeipmi.plugin/freeipmi_plugin.c)174
-rw-r--r--src/collectors/freeipmi.plugin/integrations/intelligent_platform_management_interface_ipmi.md (renamed from collectors/freeipmi.plugin/integrations/intelligent_platform_management_interface_ipmi.md)10
l---------src/collectors/idlejitter.plugin/README.md (renamed from collectors/idlejitter.plugin/README.md)0
-rw-r--r--src/collectors/idlejitter.plugin/integrations/idle_os_jitter.md (renamed from collectors/idlejitter.plugin/integrations/idle_os_jitter.md)8
-rw-r--r--src/collectors/idlejitter.plugin/metadata.yaml (renamed from collectors/idlejitter.plugin/metadata.yaml)0
-rw-r--r--src/collectors/idlejitter.plugin/plugin_idlejitter.c (renamed from collectors/idlejitter.plugin/plugin_idlejitter.c)4
l---------src/collectors/ioping.plugin/README.md (renamed from collectors/ioping.plugin/README.md)0
-rw-r--r--src/collectors/ioping.plugin/integrations/ioping.md (renamed from collectors/ioping.plugin/integrations/ioping.md)10
-rw-r--r--src/collectors/ioping.plugin/ioping.conf (renamed from collectors/ioping.plugin/ioping.conf)0
-rwxr-xr-xsrc/collectors/ioping.plugin/ioping.plugin.in (renamed from collectors/ioping.plugin/ioping.plugin.in)0
-rw-r--r--src/collectors/log2journal/log2journal-help.c (renamed from collectors/log2journal/log2journal-help.c)2
-rw-r--r--src/collectors/log2journal/log2journal-inject.c (renamed from collectors/log2journal/log2journal-inject.c)0
-rw-r--r--src/collectors/log2journal/log2journal-json.c (renamed from collectors/log2journal/log2journal-json.c)30
-rw-r--r--src/collectors/log2journal/log2journal-logfmt.c (renamed from collectors/log2journal/log2journal-logfmt.c)8
-rw-r--r--src/collectors/log2journal/log2journal-params.c (renamed from collectors/log2journal/log2journal-params.c)0
-rw-r--r--src/collectors/log2journal/log2journal-pattern.c (renamed from collectors/log2journal/log2journal-pattern.c)0
-rw-r--r--src/collectors/log2journal/log2journal-pcre2.c (renamed from collectors/log2journal/log2journal-pcre2.c)0
-rw-r--r--src/collectors/log2journal/log2journal-rename.c (renamed from collectors/log2journal/log2journal-rename.c)0
-rw-r--r--src/collectors/log2journal/log2journal-replace.c (renamed from collectors/log2journal/log2journal-replace.c)2
-rw-r--r--src/collectors/log2journal/log2journal-rewrite.c (renamed from collectors/log2journal/log2journal-rewrite.c)0
-rw-r--r--src/collectors/log2journal/log2journal-yaml.c (renamed from collectors/log2journal/log2journal-yaml.c)8
-rw-r--r--src/collectors/log2journal/log2journal.c (renamed from collectors/log2journal/log2journal.c)2
-rw-r--r--src/collectors/log2journal/log2journal.d/default.yaml (renamed from collectors/log2journal/log2journal.d/default.yaml)0
-rw-r--r--src/collectors/log2journal/log2journal.d/nginx-combined.yaml (renamed from collectors/log2journal/log2journal.d/nginx-combined.yaml)0
-rw-r--r--src/collectors/log2journal/log2journal.d/nginx-json.yaml (renamed from collectors/log2journal/log2journal.d/nginx-json.yaml)0
-rw-r--r--src/collectors/log2journal/log2journal.h (renamed from collectors/log2journal/log2journal.h)11
-rw-r--r--src/collectors/log2journal/tests.d/default.output (renamed from collectors/log2journal/tests.d/default.output)0
-rw-r--r--src/collectors/log2journal/tests.d/full.output (renamed from collectors/log2journal/tests.d/full.output)0
-rw-r--r--src/collectors/log2journal/tests.d/full.yaml (renamed from collectors/log2journal/tests.d/full.yaml)0
-rw-r--r--src/collectors/log2journal/tests.d/json-exclude.output (renamed from collectors/log2journal/tests.d/json-exclude.output)0
-rw-r--r--src/collectors/log2journal/tests.d/json-include.output (renamed from collectors/log2journal/tests.d/json-include.output)0
-rw-r--r--src/collectors/log2journal/tests.d/json.log (renamed from collectors/log2journal/tests.d/json.log)0
-rw-r--r--src/collectors/log2journal/tests.d/json.output (renamed from collectors/log2journal/tests.d/json.output)0
-rw-r--r--src/collectors/log2journal/tests.d/logfmt.log (renamed from collectors/log2journal/tests.d/logfmt.log)0
-rw-r--r--src/collectors/log2journal/tests.d/logfmt.output (renamed from collectors/log2journal/tests.d/logfmt.output)0
-rw-r--r--src/collectors/log2journal/tests.d/logfmt.yaml (renamed from collectors/log2journal/tests.d/logfmt.yaml)0
-rw-r--r--src/collectors/log2journal/tests.d/nginx-combined.log (renamed from collectors/log2journal/tests.d/nginx-combined.log)0
-rw-r--r--src/collectors/log2journal/tests.d/nginx-combined.output (renamed from collectors/log2journal/tests.d/nginx-combined.output)0
-rw-r--r--src/collectors/log2journal/tests.d/nginx-json.log (renamed from collectors/log2journal/tests.d/nginx-json.log)0
-rw-r--r--src/collectors/log2journal/tests.d/nginx-json.output (renamed from collectors/log2journal/tests.d/nginx-json.output)0
-rwxr-xr-xsrc/collectors/log2journal/tests.sh (renamed from collectors/log2journal/tests.sh)0
l---------src/collectors/macos.plugin/README.md (renamed from collectors/macos.plugin/README.md)0
-rw-r--r--src/collectors/macos.plugin/integrations/macos.md (renamed from collectors/macos.plugin/integrations/macos.md)10
-rw-r--r--src/collectors/macos.plugin/macos_fw.c (renamed from collectors/macos.plugin/macos_fw.c)0
-rw-r--r--src/collectors/macos.plugin/macos_mach_smi.c (renamed from collectors/macos.plugin/macos_mach_smi.c)0
-rw-r--r--src/collectors/macos.plugin/macos_sysctl.c (renamed from collectors/macos.plugin/macos_sysctl.c)0
-rw-r--r--src/collectors/macos.plugin/plugin_macos.c (renamed from collectors/macos.plugin/plugin_macos.c)0
-rw-r--r--src/collectors/macos.plugin/plugin_macos.h (renamed from collectors/macos.plugin/plugin_macos.h)0
l---------src/collectors/nfacct.plugin/README.md (renamed from collectors/nfacct.plugin/README.md)0
-rw-r--r--src/collectors/nfacct.plugin/integrations/netfilter.md (renamed from collectors/nfacct.plugin/integrations/netfilter.md)8
-rw-r--r--src/collectors/nfacct.plugin/metadata.yaml (renamed from collectors/nfacct.plugin/metadata.yaml)0
-rw-r--r--src/collectors/nfacct.plugin/plugin_nfacct.c (renamed from collectors/nfacct.plugin/plugin_nfacct.c)2
l---------src/collectors/perf.plugin/README.md (renamed from collectors/perf.plugin/README.md)0
-rw-r--r--src/collectors/perf.plugin/integrations/cpu_performance.md (renamed from collectors/perf.plugin/integrations/cpu_performance.md)12
-rw-r--r--src/collectors/perf.plugin/perf_plugin.c (renamed from collectors/perf.plugin/perf_plugin.c)6
-rw-r--r--src/collectors/plugins.d/plugins_d.c (renamed from collectors/plugins.d/plugins_d.c)94
-rw-r--r--src/collectors/plugins.d/pluginsd_parser.h (renamed from collectors/plugins.d/pluginsd_parser.h)17
-rw-r--r--src/collectors/proc.plugin/integrations/amd_gpu.md (renamed from collectors/proc.plugin/integrations/amd_gpu.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/conntrack.md (renamed from collectors/proc.plugin/integrations/conntrack.md)8
-rw-r--r--src/collectors/proc.plugin/integrations/disk_statistics.md (renamed from collectors/proc.plugin/integrations/disk_statistics.md)14
-rw-r--r--src/collectors/proc.plugin/integrations/entropy.md (renamed from collectors/proc.plugin/integrations/entropy.md)8
-rw-r--r--src/collectors/proc.plugin/integrations/infiniband.md (renamed from collectors/proc.plugin/integrations/infiniband.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/inter_process_communication.md (renamed from collectors/proc.plugin/integrations/inter_process_communication.md)10
-rw-r--r--src/collectors/proc.plugin/integrations/interrupts.md (renamed from collectors/proc.plugin/integrations/interrupts.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/ip_virtual_server.md (renamed from collectors/proc.plugin/integrations/ip_virtual_server.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/ipv6_socket_statistics.md (renamed from collectors/proc.plugin/integrations/ipv6_socket_statistics.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/kernel_same-page_merging.md (renamed from collectors/proc.plugin/integrations/kernel_same-page_merging.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/md_raid.md (renamed from collectors/proc.plugin/integrations/md_raid.md)14
-rw-r--r--src/collectors/proc.plugin/integrations/memory_modules_dimms.md (renamed from collectors/proc.plugin/integrations/memory_modules_dimms.md)18
-rw-r--r--src/collectors/proc.plugin/integrations/memory_statistics.md (renamed from collectors/proc.plugin/integrations/memory_statistics.md)10
-rw-r--r--src/collectors/proc.plugin/integrations/memory_usage.md (renamed from collectors/proc.plugin/integrations/memory_usage.md)14
-rw-r--r--src/collectors/proc.plugin/integrations/network_statistics.md (renamed from collectors/proc.plugin/integrations/network_statistics.md)28
-rw-r--r--src/collectors/proc.plugin/integrations/nfs_client.md (renamed from collectors/proc.plugin/integrations/nfs_client.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/nfs_server.md (renamed from collectors/proc.plugin/integrations/nfs_server.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/non-uniform_memory_access.md (renamed from collectors/proc.plugin/integrations/non-uniform_memory_access.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/page_types.md (renamed from collectors/proc.plugin/integrations/page_types.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/power_supply.md (renamed from collectors/proc.plugin/integrations/power_supply.md)8
-rw-r--r--src/collectors/proc.plugin/integrations/pressure_stall_information.md (renamed from collectors/proc.plugin/integrations/pressure_stall_information.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/sctp_statistics.md (renamed from collectors/proc.plugin/integrations/sctp_statistics.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/socket_statistics.md (renamed from collectors/proc.plugin/integrations/socket_statistics.md)10
-rw-r--r--src/collectors/proc.plugin/integrations/softirq_statistics.md (renamed from collectors/proc.plugin/integrations/softirq_statistics.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/softnet_statistics.md (renamed from collectors/proc.plugin/integrations/softnet_statistics.md)10
-rw-r--r--src/collectors/proc.plugin/integrations/synproxy.md (renamed from collectors/proc.plugin/integrations/synproxy.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/system_load_average.md (renamed from collectors/proc.plugin/integrations/system_load_average.md)16
-rw-r--r--src/collectors/proc.plugin/integrations/system_statistics.md (renamed from collectors/proc.plugin/integrations/system_statistics.md)14
-rw-r--r--src/collectors/proc.plugin/integrations/system_uptime.md (renamed from collectors/proc.plugin/integrations/system_uptime.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/wireless_network_interfaces.md (renamed from collectors/proc.plugin/integrations/wireless_network_interfaces.md)6
-rw-r--r--src/collectors/proc.plugin/integrations/zfs_adaptive_replacement_cache.md (renamed from collectors/proc.plugin/integrations/zfs_adaptive_replacement_cache.md)8
-rw-r--r--src/collectors/proc.plugin/integrations/zfs_pools.md (renamed from collectors/proc.plugin/integrations/zfs_pools.md)10
-rw-r--r--src/collectors/proc.plugin/integrations/zram.md (renamed from collectors/proc.plugin/integrations/zram.md)6
-rw-r--r--src/collectors/proc.plugin/ipc.c (renamed from collectors/proc.plugin/ipc.c)10
-rw-r--r--src/collectors/proc.plugin/plugin_proc.c (renamed from collectors/proc.plugin/plugin_proc.c)86
-rw-r--r--src/collectors/proc.plugin/plugin_proc.h (renamed from collectors/proc.plugin/plugin_proc.h)4
-rw-r--r--src/collectors/proc.plugin/proc_diskstats.c (renamed from collectors/proc.plugin/proc_diskstats.c)41
-rw-r--r--src/collectors/proc.plugin/proc_interrupts.c (renamed from collectors/proc.plugin/proc_interrupts.c)2
-rw-r--r--src/collectors/proc.plugin/proc_loadavg.c (renamed from collectors/proc.plugin/proc_loadavg.c)6
-rw-r--r--src/collectors/proc.plugin/proc_mdstat.c (renamed from collectors/proc.plugin/proc_mdstat.c)0
-rw-r--r--src/collectors/proc.plugin/proc_meminfo.c (renamed from collectors/proc.plugin/proc_meminfo.c)0
-rw-r--r--src/collectors/proc.plugin/proc_net_dev.c (renamed from collectors/proc.plugin/proc_net_dev.c)386
-rw-r--r--src/collectors/proc.plugin/proc_net_ip_vs_stats.c (renamed from collectors/proc.plugin/proc_net_ip_vs_stats.c)0
-rw-r--r--src/collectors/proc.plugin/proc_net_netstat.c (renamed from collectors/proc.plugin/proc_net_netstat.c)4
-rw-r--r--src/collectors/proc.plugin/proc_net_rpc_nfs.c (renamed from collectors/proc.plugin/proc_net_rpc_nfs.c)0
-rw-r--r--src/collectors/proc.plugin/proc_net_rpc_nfsd.c (renamed from collectors/proc.plugin/proc_net_rpc_nfsd.c)0
-rw-r--r--src/collectors/proc.plugin/proc_net_sctp_snmp.c (renamed from collectors/proc.plugin/proc_net_sctp_snmp.c)0
-rw-r--r--src/collectors/proc.plugin/proc_net_sockstat.c (renamed from collectors/proc.plugin/proc_net_sockstat.c)18
-rw-r--r--src/collectors/proc.plugin/proc_net_sockstat6.c (renamed from collectors/proc.plugin/proc_net_sockstat6.c)0
-rw-r--r--src/collectors/proc.plugin/proc_net_softnet_stat.c (renamed from collectors/proc.plugin/proc_net_softnet_stat.c)5
-rw-r--r--src/collectors/proc.plugin/proc_net_stat_conntrack.c (renamed from collectors/proc.plugin/proc_net_stat_conntrack.c)4
-rw-r--r--src/collectors/proc.plugin/proc_net_stat_synproxy.c (renamed from collectors/proc.plugin/proc_net_stat_synproxy.c)0
-rw-r--r--src/collectors/proc.plugin/proc_net_wireless.c (renamed from collectors/proc.plugin/proc_net_wireless.c)0
-rw-r--r--src/collectors/proc.plugin/proc_pagetypeinfo.c (renamed from collectors/proc.plugin/proc_pagetypeinfo.c)0
-rw-r--r--src/collectors/proc.plugin/proc_pressure.c (renamed from collectors/proc.plugin/proc_pressure.c)0
-rw-r--r--src/collectors/proc.plugin/proc_pressure.h (renamed from collectors/proc.plugin/proc_pressure.h)0
-rw-r--r--src/collectors/proc.plugin/proc_self_mountinfo.c (renamed from collectors/proc.plugin/proc_self_mountinfo.c)0
-rw-r--r--src/collectors/proc.plugin/proc_self_mountinfo.h (renamed from collectors/proc.plugin/proc_self_mountinfo.h)0
-rw-r--r--src/collectors/proc.plugin/proc_softirqs.c (renamed from collectors/proc.plugin/proc_softirqs.c)2
-rw-r--r--src/collectors/proc.plugin/proc_spl_kstat_zfs.c (renamed from collectors/proc.plugin/proc_spl_kstat_zfs.c)8
-rw-r--r--src/collectors/proc.plugin/proc_stat.c (renamed from collectors/proc.plugin/proc_stat.c)131
-rw-r--r--src/collectors/proc.plugin/proc_sys_fs_file_nr.c (renamed from collectors/proc.plugin/proc_sys_fs_file_nr.c)0
-rw-r--r--src/collectors/proc.plugin/proc_sys_kernel_random_entropy_avail.c (renamed from collectors/proc.plugin/proc_sys_kernel_random_entropy_avail.c)0
-rw-r--r--src/collectors/proc.plugin/proc_uptime.c (renamed from collectors/proc.plugin/proc_uptime.c)0
-rw-r--r--src/collectors/proc.plugin/proc_vmstat.c (renamed from collectors/proc.plugin/proc_vmstat.c)0
-rw-r--r--src/collectors/proc.plugin/sys_block_zram.c (renamed from collectors/proc.plugin/sys_block_zram.c)0
-rw-r--r--src/collectors/proc.plugin/sys_class_drm.c (renamed from collectors/proc.plugin/sys_class_drm.c)6
-rw-r--r--src/collectors/proc.plugin/sys_class_infiniband.c (renamed from collectors/proc.plugin/sys_class_infiniband.c)8
-rw-r--r--src/collectors/proc.plugin/sys_class_power_supply.c (renamed from collectors/proc.plugin/sys_class_power_supply.c)4
-rw-r--r--src/collectors/proc.plugin/sys_devices_pci_aer.c (renamed from collectors/proc.plugin/sys_devices_pci_aer.c)0
-rw-r--r--src/collectors/proc.plugin/sys_devices_system_edac_mc.c (renamed from collectors/proc.plugin/sys_devices_system_edac_mc.c)24
-rw-r--r--src/collectors/proc.plugin/sys_devices_system_node.c (renamed from collectors/proc.plugin/sys_devices_system_node.c)0
-rw-r--r--src/collectors/proc.plugin/sys_fs_btrfs.c (renamed from collectors/proc.plugin/sys_fs_btrfs.c)6
-rw-r--r--src/collectors/proc.plugin/sys_kernel_mm_ksm.c (renamed from collectors/proc.plugin/sys_kernel_mm_ksm.c)2
-rw-r--r--src/collectors/proc.plugin/zfs_common.c (renamed from collectors/proc.plugin/zfs_common.c)0
-rw-r--r--src/collectors/proc.plugin/zfs_common.h (renamed from collectors/proc.plugin/zfs_common.h)0
-rw-r--r--src/collectors/profile.plugin/README.md (renamed from collectors/profile.plugin/README.md)0
-rw-r--r--src/collectors/profile.plugin/plugin_profile.cc (renamed from collectors/profile.plugin/plugin_profile.cc)2
l---------src/collectors/python.d.plugin/adaptec_raid/README.md (renamed from collectors/python.d.plugin/adaptec_raid/README.md)0
-rw-r--r--src/collectors/python.d.plugin/adaptec_raid/adaptec_raid.chart.py (renamed from collectors/python.d.plugin/adaptec_raid/adaptec_raid.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/adaptec_raid/adaptec_raid.conf (renamed from collectors/python.d.plugin/adaptec_raid/adaptec_raid.conf)0
-rw-r--r--src/collectors/python.d.plugin/adaptec_raid/integrations/adaptecraid.md (renamed from collectors/python.d.plugin/adaptec_raid/integrations/adaptecraid.md)12
l---------src/collectors/python.d.plugin/alarms/README.md (renamed from collectors/python.d.plugin/alarms/README.md)0
-rw-r--r--src/collectors/python.d.plugin/alarms/alarms.chart.py (renamed from collectors/python.d.plugin/alarms/alarms.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/alarms/alarms.conf (renamed from collectors/python.d.plugin/alarms/alarms.conf)0
-rw-r--r--src/collectors/python.d.plugin/alarms/integrations/netdata_agent_alarms.md (renamed from collectors/python.d.plugin/alarms/integrations/netdata_agent_alarms.md)8
l---------src/collectors/python.d.plugin/am2320/README.md (renamed from collectors/python.d.plugin/am2320/README.md)0
-rw-r--r--src/collectors/python.d.plugin/am2320/am2320.chart.py (renamed from collectors/python.d.plugin/am2320/am2320.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/am2320/am2320.conf (renamed from collectors/python.d.plugin/am2320/am2320.conf)0
-rw-r--r--src/collectors/python.d.plugin/am2320/integrations/am2320.md (renamed from collectors/python.d.plugin/am2320/integrations/am2320.md)8
-rw-r--r--src/collectors/python.d.plugin/am2320/metadata.yaml (renamed from collectors/python.d.plugin/am2320/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/anomalies/anomalies.chart.py (renamed from collectors/python.d.plugin/anomalies/anomalies.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/anomalies/anomalies.conf (renamed from collectors/python.d.plugin/anomalies/anomalies.conf)0
l---------src/collectors/python.d.plugin/beanstalk/README.md (renamed from collectors/python.d.plugin/beanstalk/README.md)0
-rw-r--r--src/collectors/python.d.plugin/beanstalk/beanstalk.chart.py (renamed from collectors/python.d.plugin/beanstalk/beanstalk.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/beanstalk/beanstalk.conf (renamed from collectors/python.d.plugin/beanstalk/beanstalk.conf)0
-rw-r--r--src/collectors/python.d.plugin/beanstalk/integrations/beanstalk.md (renamed from collectors/python.d.plugin/beanstalk/integrations/beanstalk.md)10
l---------src/collectors/python.d.plugin/bind_rndc/README.md (renamed from collectors/python.d.plugin/bind_rndc/README.md)0
-rw-r--r--src/collectors/python.d.plugin/bind_rndc/bind_rndc.chart.py (renamed from collectors/python.d.plugin/bind_rndc/bind_rndc.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/bind_rndc/bind_rndc.conf (renamed from collectors/python.d.plugin/bind_rndc/bind_rndc.conf)0
-rw-r--r--src/collectors/python.d.plugin/bind_rndc/integrations/isc_bind_rndc.md (renamed from collectors/python.d.plugin/bind_rndc/integrations/isc_bind_rndc.md)10
l---------src/collectors/python.d.plugin/boinc/README.md (renamed from collectors/python.d.plugin/boinc/README.md)0
-rw-r--r--src/collectors/python.d.plugin/boinc/boinc.chart.py (renamed from collectors/python.d.plugin/boinc/boinc.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/boinc/boinc.conf (renamed from collectors/python.d.plugin/boinc/boinc.conf)0
-rw-r--r--src/collectors/python.d.plugin/boinc/integrations/boinc.md (renamed from collectors/python.d.plugin/boinc/integrations/boinc.md)16
l---------src/collectors/python.d.plugin/ceph/README.md (renamed from collectors/python.d.plugin/ceph/README.md)0
-rw-r--r--src/collectors/python.d.plugin/ceph/ceph.chart.py (renamed from collectors/python.d.plugin/ceph/ceph.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/ceph/ceph.conf (renamed from collectors/python.d.plugin/ceph/ceph.conf)0
-rw-r--r--src/collectors/python.d.plugin/ceph/integrations/ceph.md (renamed from collectors/python.d.plugin/ceph/integrations/ceph.md)10
l---------src/collectors/python.d.plugin/changefinder/README.md (renamed from collectors/python.d.plugin/changefinder/README.md)0
-rw-r--r--src/collectors/python.d.plugin/changefinder/changefinder.chart.py (renamed from collectors/python.d.plugin/changefinder/changefinder.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/changefinder/changefinder.conf (renamed from collectors/python.d.plugin/changefinder/changefinder.conf)0
-rw-r--r--src/collectors/python.d.plugin/changefinder/integrations/python.d_changefinder.md (renamed from collectors/python.d.plugin/changefinder/integrations/python.d_changefinder.md)8
-rw-r--r--src/collectors/python.d.plugin/changefinder/metadata.yaml (renamed from collectors/python.d.plugin/changefinder/metadata.yaml)0
l---------src/collectors/python.d.plugin/dovecot/README.md (renamed from collectors/python.d.plugin/dovecot/README.md)0
-rw-r--r--src/collectors/python.d.plugin/dovecot/dovecot.chart.py (renamed from collectors/python.d.plugin/dovecot/dovecot.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/dovecot/dovecot.conf (renamed from collectors/python.d.plugin/dovecot/dovecot.conf)0
-rw-r--r--src/collectors/python.d.plugin/dovecot/integrations/dovecot.md (renamed from collectors/python.d.plugin/dovecot/integrations/dovecot.md)8
-rw-r--r--src/collectors/python.d.plugin/dovecot/metadata.yaml (renamed from collectors/python.d.plugin/dovecot/metadata.yaml)0
l---------src/collectors/python.d.plugin/example/README.md (renamed from collectors/python.d.plugin/example/README.md)0
-rw-r--r--src/collectors/python.d.plugin/example/example.chart.py (renamed from collectors/python.d.plugin/example/example.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/example/example.conf (renamed from collectors/python.d.plugin/example/example.conf)0
-rw-r--r--src/collectors/python.d.plugin/example/integrations/example_collector.md (renamed from collectors/python.d.plugin/example/integrations/example_collector.md)10
l---------src/collectors/python.d.plugin/exim/README.md (renamed from collectors/python.d.plugin/exim/README.md)0
-rw-r--r--src/collectors/python.d.plugin/exim/exim.chart.py (renamed from collectors/python.d.plugin/exim/exim.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/exim/exim.conf (renamed from collectors/python.d.plugin/exim/exim.conf)0
-rw-r--r--src/collectors/python.d.plugin/exim/integrations/exim.md (renamed from collectors/python.d.plugin/exim/integrations/exim.md)8
-rw-r--r--src/collectors/python.d.plugin/exim/metadata.yaml (renamed from collectors/python.d.plugin/exim/metadata.yaml)0
l---------src/collectors/python.d.plugin/fail2ban/README.md (renamed from collectors/python.d.plugin/fail2ban/README.md)0
-rw-r--r--src/collectors/python.d.plugin/fail2ban/fail2ban.chart.py (renamed from collectors/python.d.plugin/fail2ban/fail2ban.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/fail2ban/fail2ban.conf (renamed from collectors/python.d.plugin/fail2ban/fail2ban.conf)0
-rw-r--r--src/collectors/python.d.plugin/fail2ban/integrations/fail2ban.md (renamed from collectors/python.d.plugin/fail2ban/integrations/fail2ban.md)8
-rw-r--r--src/collectors/python.d.plugin/fail2ban/metadata.yaml (renamed from collectors/python.d.plugin/fail2ban/metadata.yaml)0
l---------src/collectors/python.d.plugin/gearman/README.md (renamed from collectors/python.d.plugin/gearman/README.md)0
-rw-r--r--src/collectors/python.d.plugin/gearman/gearman.chart.py (renamed from collectors/python.d.plugin/gearman/gearman.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/gearman/gearman.conf (renamed from collectors/python.d.plugin/gearman/gearman.conf)0
-rw-r--r--src/collectors/python.d.plugin/gearman/integrations/gearman.md (renamed from collectors/python.d.plugin/gearman/integrations/gearman.md)10
l---------src/collectors/python.d.plugin/go_expvar/README.md (renamed from collectors/python.d.plugin/go_expvar/README.md)0
-rw-r--r--src/collectors/python.d.plugin/go_expvar/go_expvar.chart.py (renamed from collectors/python.d.plugin/go_expvar/go_expvar.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/go_expvar/go_expvar.conf (renamed from collectors/python.d.plugin/go_expvar/go_expvar.conf)2
-rw-r--r--src/collectors/python.d.plugin/go_expvar/integrations/go_applications_expvar.md (renamed from collectors/python.d.plugin/go_expvar/integrations/go_applications_expvar.md)18
-rw-r--r--src/collectors/python.d.plugin/haproxy/haproxy.chart.py (renamed from collectors/python.d.plugin/haproxy/haproxy.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/haproxy/haproxy.conf (renamed from collectors/python.d.plugin/haproxy/haproxy.conf)0
l---------src/collectors/python.d.plugin/hddtemp/README.md (renamed from collectors/python.d.plugin/hddtemp/README.md)0
-rw-r--r--src/collectors/python.d.plugin/hddtemp/hddtemp.chart.py (renamed from collectors/python.d.plugin/hddtemp/hddtemp.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/hddtemp/hddtemp.conf (renamed from collectors/python.d.plugin/hddtemp/hddtemp.conf)0
-rw-r--r--src/collectors/python.d.plugin/hddtemp/integrations/hdd_temperature.md (renamed from collectors/python.d.plugin/hddtemp/integrations/hdd_temperature.md)8
-rw-r--r--src/collectors/python.d.plugin/hddtemp/metadata.yaml (renamed from collectors/python.d.plugin/hddtemp/metadata.yaml)0
l---------src/collectors/python.d.plugin/hpssa/README.md (renamed from collectors/python.d.plugin/hpssa/README.md)0
-rw-r--r--src/collectors/python.d.plugin/hpssa/hpssa.chart.py (renamed from collectors/python.d.plugin/hpssa/hpssa.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/hpssa/hpssa.conf (renamed from collectors/python.d.plugin/hpssa/hpssa.conf)0
-rw-r--r--src/collectors/python.d.plugin/hpssa/integrations/hp_smart_storage_arrays.md (renamed from collectors/python.d.plugin/hpssa/integrations/hp_smart_storage_arrays.md)12
l---------src/collectors/python.d.plugin/icecast/README.md (renamed from collectors/python.d.plugin/icecast/README.md)0
-rw-r--r--src/collectors/python.d.plugin/icecast/icecast.chart.py (renamed from collectors/python.d.plugin/icecast/icecast.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/icecast/icecast.conf (renamed from collectors/python.d.plugin/icecast/icecast.conf)0
-rw-r--r--src/collectors/python.d.plugin/icecast/integrations/icecast.md (renamed from collectors/python.d.plugin/icecast/integrations/icecast.md)8
-rw-r--r--src/collectors/python.d.plugin/icecast/metadata.yaml (renamed from collectors/python.d.plugin/icecast/metadata.yaml)0
l---------src/collectors/python.d.plugin/ipfs/README.md (renamed from collectors/python.d.plugin/ipfs/README.md)0
-rw-r--r--src/collectors/python.d.plugin/ipfs/integrations/ipfs.md (renamed from collectors/python.d.plugin/ipfs/integrations/ipfs.md)10
-rw-r--r--src/collectors/python.d.plugin/ipfs/ipfs.chart.py (renamed from collectors/python.d.plugin/ipfs/ipfs.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/ipfs/ipfs.conf (renamed from collectors/python.d.plugin/ipfs/ipfs.conf)0
l---------src/collectors/python.d.plugin/litespeed/README.md (renamed from collectors/python.d.plugin/litespeed/README.md)0
-rw-r--r--src/collectors/python.d.plugin/litespeed/integrations/litespeed.md (renamed from collectors/python.d.plugin/litespeed/integrations/litespeed.md)8
-rw-r--r--src/collectors/python.d.plugin/litespeed/litespeed.chart.py (renamed from collectors/python.d.plugin/litespeed/litespeed.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/litespeed/litespeed.conf (renamed from collectors/python.d.plugin/litespeed/litespeed.conf)0
-rw-r--r--src/collectors/python.d.plugin/litespeed/metadata.yaml (renamed from collectors/python.d.plugin/litespeed/metadata.yaml)0
l---------src/collectors/python.d.plugin/megacli/README.md (renamed from collectors/python.d.plugin/megacli/README.md)0
-rw-r--r--src/collectors/python.d.plugin/megacli/integrations/megacli.md (renamed from collectors/python.d.plugin/megacli/integrations/megacli.md)18
-rw-r--r--src/collectors/python.d.plugin/megacli/megacli.chart.py (renamed from collectors/python.d.plugin/megacli/megacli.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/megacli/megacli.conf (renamed from collectors/python.d.plugin/megacli/megacli.conf)0
l---------src/collectors/python.d.plugin/memcached/README.md (renamed from collectors/python.d.plugin/memcached/README.md)0
-rw-r--r--src/collectors/python.d.plugin/memcached/integrations/memcached.md (renamed from collectors/python.d.plugin/memcached/integrations/memcached.md)14
-rw-r--r--src/collectors/python.d.plugin/memcached/memcached.chart.py (renamed from collectors/python.d.plugin/memcached/memcached.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/memcached/memcached.conf (renamed from collectors/python.d.plugin/memcached/memcached.conf)0
l---------src/collectors/python.d.plugin/monit/README.md (renamed from collectors/python.d.plugin/monit/README.md)0
-rw-r--r--src/collectors/python.d.plugin/monit/integrations/monit.md (renamed from collectors/python.d.plugin/monit/integrations/monit.md)8
-rw-r--r--src/collectors/python.d.plugin/monit/metadata.yaml (renamed from collectors/python.d.plugin/monit/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/monit/monit.chart.py (renamed from collectors/python.d.plugin/monit/monit.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/monit/monit.conf (renamed from collectors/python.d.plugin/monit/monit.conf)0
l---------src/collectors/python.d.plugin/nsd/README.md (renamed from collectors/python.d.plugin/nsd/README.md)0
-rw-r--r--src/collectors/python.d.plugin/nsd/integrations/name_server_daemon.md (renamed from collectors/python.d.plugin/nsd/integrations/name_server_daemon.md)8
-rw-r--r--src/collectors/python.d.plugin/nsd/metadata.yaml (renamed from collectors/python.d.plugin/nsd/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/nsd/nsd.chart.py (renamed from collectors/python.d.plugin/nsd/nsd.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/nsd/nsd.conf (renamed from collectors/python.d.plugin/nsd/nsd.conf)0
-rw-r--r--src/collectors/python.d.plugin/nvidia_smi/nvidia_smi.chart.py (renamed from collectors/python.d.plugin/nvidia_smi/nvidia_smi.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/nvidia_smi/nvidia_smi.conf (renamed from collectors/python.d.plugin/nvidia_smi/nvidia_smi.conf)0
l---------src/collectors/python.d.plugin/openldap/README.md (renamed from collectors/python.d.plugin/openldap/README.md)0
-rw-r--r--src/collectors/python.d.plugin/openldap/integrations/openldap.md (renamed from collectors/python.d.plugin/openldap/integrations/openldap.md)8
-rw-r--r--src/collectors/python.d.plugin/openldap/metadata.yaml (renamed from collectors/python.d.plugin/openldap/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/openldap/openldap.chart.py (renamed from collectors/python.d.plugin/openldap/openldap.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/openldap/openldap.conf (renamed from collectors/python.d.plugin/openldap/openldap.conf)0
l---------src/collectors/python.d.plugin/oracledb/README.md (renamed from collectors/python.d.plugin/oracledb/README.md)0
-rw-r--r--src/collectors/python.d.plugin/oracledb/integrations/oracle_db.md (renamed from collectors/python.d.plugin/oracledb/integrations/oracle_db.md)8
-rw-r--r--src/collectors/python.d.plugin/oracledb/metadata.yaml (renamed from collectors/python.d.plugin/oracledb/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/oracledb/oracledb.chart.py (renamed from collectors/python.d.plugin/oracledb/oracledb.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/oracledb/oracledb.conf (renamed from collectors/python.d.plugin/oracledb/oracledb.conf)0
l---------src/collectors/python.d.plugin/pandas/README.md (renamed from collectors/python.d.plugin/pandas/README.md)0
-rw-r--r--src/collectors/python.d.plugin/pandas/integrations/pandas.md (renamed from collectors/python.d.plugin/pandas/integrations/pandas.md)8
-rw-r--r--src/collectors/python.d.plugin/pandas/metadata.yaml (renamed from collectors/python.d.plugin/pandas/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/pandas/pandas.chart.py (renamed from collectors/python.d.plugin/pandas/pandas.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/pandas/pandas.conf (renamed from collectors/python.d.plugin/pandas/pandas.conf)0
l---------src/collectors/python.d.plugin/postfix/README.md (renamed from collectors/python.d.plugin/postfix/README.md)0
-rw-r--r--src/collectors/python.d.plugin/postfix/integrations/postfix.md (renamed from collectors/python.d.plugin/postfix/integrations/postfix.md)6
-rw-r--r--src/collectors/python.d.plugin/postfix/metadata.yaml (renamed from collectors/python.d.plugin/postfix/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/postfix/postfix.chart.py (renamed from collectors/python.d.plugin/postfix/postfix.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/postfix/postfix.conf (renamed from collectors/python.d.plugin/postfix/postfix.conf)0
l---------src/collectors/python.d.plugin/puppet/README.md (renamed from collectors/python.d.plugin/puppet/README.md)0
-rw-r--r--src/collectors/python.d.plugin/puppet/integrations/puppet.md (renamed from collectors/python.d.plugin/puppet/integrations/puppet.md)8
-rw-r--r--src/collectors/python.d.plugin/puppet/metadata.yaml (renamed from collectors/python.d.plugin/puppet/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/puppet/puppet.chart.py (renamed from collectors/python.d.plugin/puppet/puppet.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/puppet/puppet.conf (renamed from collectors/python.d.plugin/puppet/puppet.conf)0
-rw-r--r--src/collectors/python.d.plugin/python.d.conf (renamed from collectors/python.d.plugin/python.d.conf)0
-rw-r--r--src/collectors/python.d.plugin/python.d.plugin.in (renamed from collectors/python.d.plugin/python.d.plugin.in)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/__init__.py (renamed from collectors/python.d.plugin/python_modules/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/ExecutableService.py (renamed from collectors/python.d.plugin/python_modules/bases/FrameworkServices/ExecutableService.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/LogService.py (renamed from collectors/python.d.plugin/python_modules/bases/FrameworkServices/LogService.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/MySQLService.py (renamed from collectors/python.d.plugin/python_modules/bases/FrameworkServices/MySQLService.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/SimpleService.py (renamed from collectors/python.d.plugin/python_modules/bases/FrameworkServices/SimpleService.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/SocketService.py (renamed from collectors/python.d.plugin/python_modules/bases/FrameworkServices/SocketService.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/UrlService.py (renamed from collectors/python.d.plugin/python_modules/bases/FrameworkServices/UrlService.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/__init__.py (renamed from collectors/python.d.plugin/python_modules/bases/FrameworkServices/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/__init__.py (renamed from collectors/python.d.plugin/python_modules/bases/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/charts.py (renamed from collectors/python.d.plugin/python_modules/bases/charts.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/collection.py (renamed from collectors/python.d.plugin/python_modules/bases/collection.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/loaders.py (renamed from collectors/python.d.plugin/python_modules/bases/loaders.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/bases/loggers.py (renamed from collectors/python.d.plugin/python_modules/bases/loggers.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/__init__.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/composer.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/composer.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/constructor.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/constructor.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/cyaml.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/cyaml.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/dumper.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/dumper.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/emitter.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/emitter.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/error.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/error.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/events.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/events.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/loader.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/loader.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/nodes.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/nodes.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/parser.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/parser.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/reader.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/reader.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/representer.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/representer.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/resolver.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/resolver.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/scanner.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/scanner.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/serializer.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/serializer.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml2/tokens.py (renamed from collectors/python.d.plugin/python_modules/pyyaml2/tokens.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/__init__.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/composer.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/composer.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/constructor.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/constructor.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/cyaml.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/cyaml.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/dumper.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/dumper.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/emitter.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/emitter.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/error.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/error.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/events.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/events.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/loader.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/loader.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/nodes.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/nodes.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/parser.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/parser.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/reader.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/reader.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/representer.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/representer.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/resolver.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/resolver.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/scanner.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/scanner.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/serializer.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/serializer.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/pyyaml3/tokens.py (renamed from collectors/python.d.plugin/python_modules/pyyaml3/tokens.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/third_party/__init__.py (renamed from collectors/python.d.plugin/python_modules/third_party/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/third_party/boinc_client.py (renamed from collectors/python.d.plugin/python_modules/third_party/boinc_client.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/third_party/filelock.py (renamed from collectors/python.d.plugin/python_modules/third_party/filelock.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/third_party/lm_sensors.py (renamed from collectors/python.d.plugin/python_modules/third_party/lm_sensors.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/third_party/mcrcon.py (renamed from collectors/python.d.plugin/python_modules/third_party/mcrcon.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/third_party/monotonic.py (renamed from collectors/python.d.plugin/python_modules/third_party/monotonic.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/third_party/ordereddict.py (renamed from collectors/python.d.plugin/python_modules/third_party/ordereddict.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/__init__.py (renamed from collectors/python.d.plugin/python_modules/urllib3/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/_collections.py (renamed from collectors/python.d.plugin/python_modules/urllib3/_collections.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/connection.py (renamed from collectors/python.d.plugin/python_modules/urllib3/connection.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/connectionpool.py (renamed from collectors/python.d.plugin/python_modules/urllib3/connectionpool.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/contrib/__init__.py (renamed from collectors/python.d.plugin/python_modules/urllib3/contrib/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/__init__.py (renamed from collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/bindings.py (renamed from collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/bindings.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/low_level.py (renamed from collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/low_level.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/contrib/appengine.py (renamed from collectors/python.d.plugin/python_modules/urllib3/contrib/appengine.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/contrib/ntlmpool.py (renamed from collectors/python.d.plugin/python_modules/urllib3/contrib/ntlmpool.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/contrib/pyopenssl.py (renamed from collectors/python.d.plugin/python_modules/urllib3/contrib/pyopenssl.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/contrib/securetransport.py (renamed from collectors/python.d.plugin/python_modules/urllib3/contrib/securetransport.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/contrib/socks.py (renamed from collectors/python.d.plugin/python_modules/urllib3/contrib/socks.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/exceptions.py (renamed from collectors/python.d.plugin/python_modules/urllib3/exceptions.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/fields.py (renamed from collectors/python.d.plugin/python_modules/urllib3/fields.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/filepost.py (renamed from collectors/python.d.plugin/python_modules/urllib3/filepost.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/packages/__init__.py (renamed from collectors/python.d.plugin/python_modules/urllib3/packages/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/packages/backports/__init__.py (renamed from collectors/python.d.plugin/python_modules/urllib3/packages/backports/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/packages/backports/makefile.py (renamed from collectors/python.d.plugin/python_modules/urllib3/packages/backports/makefile.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/packages/ordered_dict.py (renamed from collectors/python.d.plugin/python_modules/urllib3/packages/ordered_dict.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/packages/six.py (renamed from collectors/python.d.plugin/python_modules/urllib3/packages/six.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/__init__.py (renamed from collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/_implementation.py (renamed from collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/_implementation.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/poolmanager.py (renamed from collectors/python.d.plugin/python_modules/urllib3/poolmanager.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/request.py (renamed from collectors/python.d.plugin/python_modules/urllib3/request.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/response.py (renamed from collectors/python.d.plugin/python_modules/urllib3/response.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/util/__init__.py (renamed from collectors/python.d.plugin/python_modules/urllib3/util/__init__.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/util/connection.py (renamed from collectors/python.d.plugin/python_modules/urllib3/util/connection.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/util/request.py (renamed from collectors/python.d.plugin/python_modules/urllib3/util/request.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/util/response.py (renamed from collectors/python.d.plugin/python_modules/urllib3/util/response.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/util/retry.py (renamed from collectors/python.d.plugin/python_modules/urllib3/util/retry.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/util/selectors.py (renamed from collectors/python.d.plugin/python_modules/urllib3/util/selectors.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/util/ssl_.py (renamed from collectors/python.d.plugin/python_modules/urllib3/util/ssl_.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/util/timeout.py (renamed from collectors/python.d.plugin/python_modules/urllib3/util/timeout.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/util/url.py (renamed from collectors/python.d.plugin/python_modules/urllib3/util/url.py)0
-rw-r--r--src/collectors/python.d.plugin/python_modules/urllib3/util/wait.py (renamed from collectors/python.d.plugin/python_modules/urllib3/util/wait.py)0
l---------src/collectors/python.d.plugin/rethinkdbs/README.md (renamed from collectors/python.d.plugin/rethinkdbs/README.md)0
-rw-r--r--src/collectors/python.d.plugin/rethinkdbs/integrations/rethinkdb.md (renamed from collectors/python.d.plugin/rethinkdbs/integrations/rethinkdb.md)8
-rw-r--r--src/collectors/python.d.plugin/rethinkdbs/metadata.yaml (renamed from collectors/python.d.plugin/rethinkdbs/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/rethinkdbs/rethinkdbs.chart.py (renamed from collectors/python.d.plugin/rethinkdbs/rethinkdbs.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/rethinkdbs/rethinkdbs.conf (renamed from collectors/python.d.plugin/rethinkdbs/rethinkdbs.conf)0
l---------src/collectors/python.d.plugin/retroshare/README.md (renamed from collectors/python.d.plugin/retroshare/README.md)0
-rw-r--r--src/collectors/python.d.plugin/retroshare/integrations/retroshare.md (renamed from collectors/python.d.plugin/retroshare/integrations/retroshare.md)10
-rw-r--r--src/collectors/python.d.plugin/retroshare/retroshare.chart.py (renamed from collectors/python.d.plugin/retroshare/retroshare.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/retroshare/retroshare.conf (renamed from collectors/python.d.plugin/retroshare/retroshare.conf)0
l---------src/collectors/python.d.plugin/riakkv/README.md (renamed from collectors/python.d.plugin/riakkv/README.md)0
-rw-r--r--src/collectors/python.d.plugin/riakkv/integrations/riakkv.md (renamed from collectors/python.d.plugin/riakkv/integrations/riakkv.md)20
-rw-r--r--src/collectors/python.d.plugin/riakkv/riakkv.chart.py (renamed from collectors/python.d.plugin/riakkv/riakkv.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/riakkv/riakkv.conf (renamed from collectors/python.d.plugin/riakkv/riakkv.conf)0
l---------src/collectors/python.d.plugin/samba/README.md (renamed from collectors/python.d.plugin/samba/README.md)0
-rw-r--r--src/collectors/python.d.plugin/samba/integrations/samba.md (renamed from collectors/python.d.plugin/samba/integrations/samba.md)12
-rw-r--r--src/collectors/python.d.plugin/samba/samba.chart.py (renamed from collectors/python.d.plugin/samba/samba.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/samba/samba.conf (renamed from collectors/python.d.plugin/samba/samba.conf)0
l---------src/collectors/python.d.plugin/sensors/README.md (renamed from collectors/python.d.plugin/sensors/README.md)0
-rw-r--r--src/collectors/python.d.plugin/sensors/integrations/linux_sensors_lm-sensors.md (renamed from collectors/python.d.plugin/sensors/integrations/linux_sensors_lm-sensors.md)8
-rw-r--r--src/collectors/python.d.plugin/sensors/sensors.chart.py (renamed from collectors/python.d.plugin/sensors/sensors.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/sensors/sensors.conf (renamed from collectors/python.d.plugin/sensors/sensors.conf)0
l---------src/collectors/python.d.plugin/smartd_log/README.md (renamed from collectors/python.d.plugin/smartd_log/README.md)0
-rw-r--r--src/collectors/python.d.plugin/smartd_log/integrations/s.m.a.r.t..md (renamed from collectors/python.d.plugin/smartd_log/integrations/s.m.a.r.t..md)8
-rw-r--r--src/collectors/python.d.plugin/smartd_log/metadata.yaml (renamed from collectors/python.d.plugin/smartd_log/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/smartd_log/smartd_log.chart.py (renamed from collectors/python.d.plugin/smartd_log/smartd_log.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/smartd_log/smartd_log.conf (renamed from collectors/python.d.plugin/smartd_log/smartd_log.conf)0
l---------src/collectors/python.d.plugin/spigotmc/README.md (renamed from collectors/python.d.plugin/spigotmc/README.md)0
-rw-r--r--src/collectors/python.d.plugin/spigotmc/integrations/spigotmc.md (renamed from collectors/python.d.plugin/spigotmc/integrations/spigotmc.md)8
-rw-r--r--src/collectors/python.d.plugin/spigotmc/metadata.yaml (renamed from collectors/python.d.plugin/spigotmc/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/spigotmc/spigotmc.chart.py (renamed from collectors/python.d.plugin/spigotmc/spigotmc.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/spigotmc/spigotmc.conf (renamed from collectors/python.d.plugin/spigotmc/spigotmc.conf)0
l---------src/collectors/python.d.plugin/squid/README.md (renamed from collectors/python.d.plugin/squid/README.md)0
-rw-r--r--src/collectors/python.d.plugin/squid/integrations/squid.md (renamed from collectors/python.d.plugin/squid/integrations/squid.md)8
-rw-r--r--src/collectors/python.d.plugin/squid/metadata.yaml (renamed from collectors/python.d.plugin/squid/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/squid/squid.chart.py (renamed from collectors/python.d.plugin/squid/squid.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/squid/squid.conf (renamed from collectors/python.d.plugin/squid/squid.conf)0
l---------src/collectors/python.d.plugin/tomcat/README.md (renamed from collectors/python.d.plugin/tomcat/README.md)0
-rw-r--r--src/collectors/python.d.plugin/tomcat/integrations/tomcat.md (renamed from collectors/python.d.plugin/tomcat/integrations/tomcat.md)8
-rw-r--r--src/collectors/python.d.plugin/tomcat/metadata.yaml (renamed from collectors/python.d.plugin/tomcat/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/tomcat/tomcat.chart.py (renamed from collectors/python.d.plugin/tomcat/tomcat.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/tomcat/tomcat.conf (renamed from collectors/python.d.plugin/tomcat/tomcat.conf)0
l---------src/collectors/python.d.plugin/tor/README.md (renamed from collectors/python.d.plugin/tor/README.md)0
-rw-r--r--src/collectors/python.d.plugin/tor/integrations/tor.md (renamed from collectors/python.d.plugin/tor/integrations/tor.md)8
-rw-r--r--src/collectors/python.d.plugin/tor/metadata.yaml (renamed from collectors/python.d.plugin/tor/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/tor/tor.chart.py (renamed from collectors/python.d.plugin/tor/tor.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/tor/tor.conf (renamed from collectors/python.d.plugin/tor/tor.conf)0
-rw-r--r--src/collectors/python.d.plugin/traefik/traefik.chart.py (renamed from collectors/python.d.plugin/traefik/traefik.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/traefik/traefik.conf (renamed from collectors/python.d.plugin/traefik/traefik.conf)0
l---------src/collectors/python.d.plugin/uwsgi/README.md (renamed from collectors/python.d.plugin/uwsgi/README.md)0
-rw-r--r--src/collectors/python.d.plugin/uwsgi/integrations/uwsgi.md (renamed from collectors/python.d.plugin/uwsgi/integrations/uwsgi.md)8
-rw-r--r--src/collectors/python.d.plugin/uwsgi/metadata.yaml (renamed from collectors/python.d.plugin/uwsgi/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/uwsgi/uwsgi.chart.py (renamed from collectors/python.d.plugin/uwsgi/uwsgi.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/uwsgi/uwsgi.conf (renamed from collectors/python.d.plugin/uwsgi/uwsgi.conf)0
l---------src/collectors/python.d.plugin/varnish/README.md (renamed from collectors/python.d.plugin/varnish/README.md)0
-rw-r--r--src/collectors/python.d.plugin/varnish/integrations/varnish.md (renamed from collectors/python.d.plugin/varnish/integrations/varnish.md)8
-rw-r--r--src/collectors/python.d.plugin/varnish/metadata.yaml (renamed from collectors/python.d.plugin/varnish/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/varnish/varnish.chart.py (renamed from collectors/python.d.plugin/varnish/varnish.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/varnish/varnish.conf (renamed from collectors/python.d.plugin/varnish/varnish.conf)0
l---------src/collectors/python.d.plugin/w1sensor/README.md (renamed from collectors/python.d.plugin/w1sensor/README.md)0
-rw-r--r--src/collectors/python.d.plugin/w1sensor/integrations/1-wire_sensors.md (renamed from collectors/python.d.plugin/w1sensor/integrations/1-wire_sensors.md)8
-rw-r--r--src/collectors/python.d.plugin/w1sensor/metadata.yaml (renamed from collectors/python.d.plugin/w1sensor/metadata.yaml)0
-rw-r--r--src/collectors/python.d.plugin/w1sensor/w1sensor.chart.py (renamed from collectors/python.d.plugin/w1sensor/w1sensor.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/w1sensor/w1sensor.conf (renamed from collectors/python.d.plugin/w1sensor/w1sensor.conf)0
l---------src/collectors/python.d.plugin/zscores/README.md (renamed from collectors/python.d.plugin/zscores/README.md)0
-rw-r--r--src/collectors/python.d.plugin/zscores/integrations/python.d_zscores.md (renamed from collectors/python.d.plugin/zscores/integrations/python.d_zscores.md)10
-rw-r--r--src/collectors/python.d.plugin/zscores/zscores.chart.py (renamed from collectors/python.d.plugin/zscores/zscores.chart.py)0
-rw-r--r--src/collectors/python.d.plugin/zscores/zscores.conf (renamed from collectors/python.d.plugin/zscores/zscores.conf)0
l---------src/collectors/slabinfo.plugin/README.md (renamed from collectors/slabinfo.plugin/README.md)0
-rw-r--r--src/collectors/slabinfo.plugin/integrations/linux_kernel_slab_allocator_statistics.md (renamed from collectors/slabinfo.plugin/integrations/linux_kernel_slab_allocator_statistics.md)8
-rw-r--r--src/collectors/slabinfo.plugin/metadata.yaml (renamed from collectors/slabinfo.plugin/metadata.yaml)0
-rw-r--r--src/collectors/slabinfo.plugin/slabinfo.c (renamed from collectors/slabinfo.plugin/slabinfo.c)11
-rw-r--r--src/collectors/statsd.plugin/asterisk.conf (renamed from collectors/statsd.plugin/asterisk.conf)0
-rw-r--r--src/collectors/statsd.plugin/asterisk.md (renamed from collectors/statsd.plugin/asterisk.md)4
-rw-r--r--src/collectors/statsd.plugin/example.conf (renamed from collectors/statsd.plugin/example.conf)0
-rw-r--r--src/collectors/statsd.plugin/k6.conf (renamed from collectors/statsd.plugin/k6.conf)0
-rw-r--r--src/collectors/statsd.plugin/k6.md (renamed from collectors/statsd.plugin/k6.md)2
-rw-r--r--src/collectors/statsd.plugin/statsd.c (renamed from collectors/statsd.plugin/statsd.c)3
-rw-r--r--src/collectors/systemd-journal.plugin/forward_secure_sealing.md (renamed from collectors/systemd-journal.plugin/forward_secure_sealing.md)0
-rw-r--r--src/collectors/systemd-journal.plugin/passive_journal_centralization_guide_self_signed_certs.md (renamed from collectors/systemd-journal.plugin/passive_journal_centralization_guide_self_signed_certs.md)4
-rw-r--r--src/collectors/systemd-journal.plugin/systemd-internals.h (renamed from collectors/systemd-journal.plugin/systemd-internals.h)26
-rw-r--r--src/collectors/systemd-journal.plugin/systemd-journal-files.c (renamed from collectors/systemd-journal.plugin/systemd-journal-files.c)154
-rw-r--r--src/collectors/systemd-journal.plugin/systemd-journal-fstat.c (renamed from collectors/systemd-journal.plugin/systemd-journal-fstat.c)0
-rwxr-xr-xsrc/collectors/systemd-journal.plugin/systemd-journal-self-signed-certs.sh (renamed from collectors/systemd-journal.plugin/systemd-journal-self-signed-certs.sh)0
-rw-r--r--src/collectors/systemd-journal.plugin/systemd-journal-watcher.c (renamed from collectors/systemd-journal.plugin/systemd-journal-watcher.c)14
-rw-r--r--src/collectors/systemd-journal.plugin/systemd-journal.c (renamed from collectors/systemd-journal.plugin/systemd-journal.c)157
-rw-r--r--src/collectors/systemd-journal.plugin/systemd-units.c (renamed from collectors/systemd-journal.plugin/systemd-units.c)33
l---------src/collectors/tc.plugin/README.md (renamed from collectors/tc.plugin/README.md)0
-rw-r--r--src/collectors/tc.plugin/integrations/tc_qos_classes.md (renamed from collectors/tc.plugin/integrations/tc_qos_classes.md)8
-rw-r--r--src/collectors/tc.plugin/metadata.yaml (renamed from collectors/tc.plugin/metadata.yaml)0
-rw-r--r--src/collectors/tc.plugin/plugin_tc.c (renamed from collectors/tc.plugin/plugin_tc.c)0
-rwxr-xr-xsrc/collectors/tc.plugin/tc-qos-helper.sh.in (renamed from collectors/tc.plugin/tc-qos-helper.sh.in)2
l---------src/collectors/timex.plugin/README.md (renamed from collectors/timex.plugin/README.md)0
-rw-r--r--src/collectors/timex.plugin/integrations/timex.md (renamed from collectors/timex.plugin/integrations/timex.md)10
-rw-r--r--src/collectors/timex.plugin/plugin_timex.c (renamed from collectors/timex.plugin/plugin_timex.c)0
l---------src/collectors/xenstat.plugin/README.md (renamed from collectors/xenstat.plugin/README.md)0
-rw-r--r--src/collectors/xenstat.plugin/integrations/xen_xcp-ng.md (renamed from collectors/xenstat.plugin/integrations/xen_xcp-ng.md)8
-rw-r--r--src/collectors/xenstat.plugin/metadata.yaml (renamed from collectors/xenstat.plugin/metadata.yaml)0
-rw-r--r--src/collectors/xenstat.plugin/xenstat_plugin.c (renamed from collectors/xenstat.plugin/xenstat_plugin.c)7
791 files changed, 4019 insertions, 49262 deletions
diff --git a/collectors/COLLECTORS.md b/collectors/COLLECTORS.md
deleted file mode 100644
index 9a3499593..000000000
--- a/collectors/COLLECTORS.md
+++ /dev/null
@@ -1,1194 +0,0 @@
-# Monitor anything with Netdata
-
-Netdata uses collectors to help you gather metrics from your favorite applications and services and view them in
-real-time, interactive charts. The following list includes collectors for both external services/applications and
-internal system metrics.
-
-Learn more
-about [how collectors work](https://github.com/netdata/netdata/blob/master/collectors/README.md), and
-then learn how to [enable or
-configure](https://github.com/netdata/netdata/blob/master/collectors/REFERENCE.md#enable-and-disable-a-specific-collection-module) any of the below collectors using the same process.
-
-Some collectors have both Go and Python versions as we continue our effort to migrate all collectors to Go. In these
-cases, _Netdata always prioritizes the Go version_, and we highly recommend you use the Go versions for the best
-experience.
-
-If you want to use a Python version of a collector, you need to
-explicitly [disable the Go version](https://github.com/netdata/netdata/blob/master/collectors/REFERENCE.md#enable-and-disable-a-specific-collection-module),
-and enable the Python version. Netdata then skips the Go version and attempts to load the Python version and its
-accompanying configuration file.
-
-## Add your application to Netdata
-
-If you don't see the app/service you'd like to monitor in this list:
-
-- If your application has a Prometheus endpoint, Netdata can monitor it! Look at our
- [generic Prometheus collector](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/README.md).
-
-- If your application is instrumented to expose [StatsD](https://blog.netdata.cloud/introduction-to-statsd/) metrics,
- see our [generic StatsD collector](https://github.com/netdata/netdata/blob/master/collectors/statsd.plugin/README.md).
-
-- If you have data in CSV, JSON, XML or other popular formats, you may be able to use our
- [generic structured data (Pandas) collector](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/pandas/README.md),
-
-- Check out our [GitHub issues](https://github.com/netdata/netdata/issues). Use the search bar to look for previous
- discussions about that collector—we may be looking for assistance from users such as yourself!
-
-- If you don't see the collector there, you can make
- a [feature request](https://github.com/netdata/netdata/issues/new/choose) on GitHub.
-
-- If you have basic software development skills, you can add your own plugin
- in [Go](https://github.com/netdata/go.d.plugin/blob/master/README.md#how-to-develop-a-collector)
- or [Python](https://github.com/netdata/netdata/blob/master/docs/guides/python-collector.md)
-
-## Available Data Collection Integrations
-<!-- AUTOGENERATED PART BY integrations/gen_doc_collector_page.py SCRIPT, DO NOT EDIT MANUALLY -->
-### APM
-
-- [Alamos FE2 server](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/alamos_fe2_server.md)
-
-- [Apache Airflow](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/apache_airflow.md)
-
-- [Apache Flink](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/apache_flink.md)
-
-- [Audisto](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/audisto.md)
-
-- [Dependency-Track](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/dependency-track.md)
-
-- [Go applications (EXPVAR)](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/go_expvar/integrations/go_applications_expvar.md)
-
-- [Google Pagespeed](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/google_pagespeed.md)
-
-- [IBM AIX systems Njmon](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ibm_aix_systems_njmon.md)
-
-- [JMX](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/jmx.md)
-
-- [Java Spring-boot 2 applications](https://github.com/netdata/go.d.plugin/blob/master/modules/springboot2/integrations/java_spring-boot_2_applications.md)
-
-- [NRPE daemon](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/nrpe_daemon.md)
-
-- [Sentry](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/sentry.md)
-
-- [Sysload](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/sysload.md)
-
-- [VSCode](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/vscode.md)
-
-- [YOURLS URL Shortener](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/yourls_url_shortener.md)
-
-- [bpftrace variables](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/bpftrace_variables.md)
-
-- [gpsd](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/gpsd.md)
-
-- [jolokia](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/jolokia.md)
-
-- [phpDaemon](https://github.com/netdata/go.d.plugin/blob/master/modules/phpdaemon/integrations/phpdaemon.md)
-
-### Authentication and Authorization
-
-- [Fail2ban](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/fail2ban/integrations/fail2ban.md)
-
-- [FreeRADIUS](https://github.com/netdata/go.d.plugin/blob/master/modules/freeradius/integrations/freeradius.md)
-
-- [HashiCorp Vault secrets](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/hashicorp_vault_secrets.md)
-
-- [LDAP](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ldap.md)
-
-- [OpenLDAP (community)](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/openldap_community.md)
-
-- [OpenLDAP](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/openldap/integrations/openldap.md)
-
-- [RADIUS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/radius.md)
-
-- [SSH](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ssh.md)
-
-- [TACACS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/tacacs.md)
-
-### Blockchain Servers
-
-- [Chia](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/chia.md)
-
-- [Crypto exchanges](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/crypto_exchanges.md)
-
-- [Cryptowatch](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cryptowatch.md)
-
-- [Energi Core Wallet](https://github.com/netdata/go.d.plugin/blob/master/modules/energid/integrations/energi_core_wallet.md)
-
-- [Go-ethereum](https://github.com/netdata/go.d.plugin/blob/master/modules/geth/integrations/go-ethereum.md)
-
-- [Helium miner (validator)](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/helium_miner_validator.md)
-
-- [IOTA full node](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/iota_full_node.md)
-
-- [Sia](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/sia.md)
-
-### CICD Platforms
-
-- [Concourse](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/concourse.md)
-
-- [GitLab Runner](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/gitlab_runner.md)
-
-- [Jenkins](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/jenkins.md)
-
-- [Puppet](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/puppet/integrations/puppet.md)
-
-### Cloud Provider Managed
-
-- [AWS EC2 Compute instances](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/aws_ec2_compute_instances.md)
-
-- [AWS EC2 Spot Instance](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/aws_ec2_spot_instance.md)
-
-- [AWS ECS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/aws_ecs.md)
-
-- [AWS Health events](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/aws_health_events.md)
-
-- [AWS Quota](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/aws_quota.md)
-
-- [AWS S3 buckets](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/aws_s3_buckets.md)
-
-- [AWS SQS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/aws_sqs.md)
-
-- [AWS instance health](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/aws_instance_health.md)
-
-- [Akamai Global Traffic Management](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/akamai_global_traffic_management.md)
-
-- [Akami Cloudmonitor](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/akami_cloudmonitor.md)
-
-- [Alibaba Cloud](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/alibaba_cloud.md)
-
-- [ArvanCloud CDN](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/arvancloud_cdn.md)
-
-- [Azure AD App passwords](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/azure_ad_app_passwords.md)
-
-- [Azure Elastic Pool SQL](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/azure_elastic_pool_sql.md)
-
-- [Azure Resources](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/azure_resources.md)
-
-- [Azure SQL](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/azure_sql.md)
-
-- [Azure Service Bus](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/azure_service_bus.md)
-
-- [Azure application](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/azure_application.md)
-
-- [BigQuery](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/bigquery.md)
-
-- [CloudWatch](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cloudwatch.md)
-
-- [Dell EMC ECS cluster](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/dell_emc_ecs_cluster.md)
-
-- [DigitalOcean](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/digitalocean.md)
-
-- [GCP GCE](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/gcp_gce.md)
-
-- [GCP Quota](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/gcp_quota.md)
-
-- [Google Cloud Platform](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/google_cloud_platform.md)
-
-- [Google Stackdriver](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/google_stackdriver.md)
-
-- [Linode](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/linode.md)
-
-- [Lustre metadata](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/lustre_metadata.md)
-
-- [Nextcloud servers](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/nextcloud_servers.md)
-
-- [OpenStack](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/openstack.md)
-
-- [Zerto](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/zerto.md)
-
-### Containers and VMs
-
-- [Containers](https://github.com/netdata/netdata/blob/master/collectors/cgroups.plugin/integrations/containers.md)
-
-- [Docker Engine](https://github.com/netdata/go.d.plugin/blob/master/modules/docker_engine/integrations/docker_engine.md)
-
-- [Docker Hub repository](https://github.com/netdata/go.d.plugin/blob/master/modules/dockerhub/integrations/docker_hub_repository.md)
-
-- [Docker](https://github.com/netdata/go.d.plugin/blob/master/modules/docker/integrations/docker.md)
-
-- [LXC Containers](https://github.com/netdata/netdata/blob/master/collectors/cgroups.plugin/integrations/lxc_containers.md)
-
-- [Libvirt Containers](https://github.com/netdata/netdata/blob/master/collectors/cgroups.plugin/integrations/libvirt_containers.md)
-
-- [NSX-T](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/nsx-t.md)
-
-- [Podman](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/podman.md)
-
-- [Proxmox Containers](https://github.com/netdata/netdata/blob/master/collectors/cgroups.plugin/integrations/proxmox_containers.md)
-
-- [Proxmox VE](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/proxmox_ve.md)
-
-- [VMware vCenter Server](https://github.com/netdata/go.d.plugin/blob/master/modules/vsphere/integrations/vmware_vcenter_server.md)
-
-- [Virtual Machines](https://github.com/netdata/netdata/blob/master/collectors/cgroups.plugin/integrations/virtual_machines.md)
-
-- [Xen XCP-ng](https://github.com/netdata/netdata/blob/master/collectors/xenstat.plugin/integrations/xen_xcp-ng.md)
-
-- [cAdvisor](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cadvisor.md)
-
-- [oVirt Containers](https://github.com/netdata/netdata/blob/master/collectors/cgroups.plugin/integrations/ovirt_containers.md)
-
-- [vCenter Server Appliance](https://github.com/netdata/go.d.plugin/blob/master/modules/vcsa/integrations/vcenter_server_appliance.md)
-
-### Databases
-
-- [4D Server](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/4d_server.md)
-
-- [AWS RDS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/aws_rds.md)
-
-- [Cassandra](https://github.com/netdata/go.d.plugin/blob/master/modules/cassandra/integrations/cassandra.md)
-
-- [ClickHouse](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/clickhouse.md)
-
-- [ClusterControl CMON](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/clustercontrol_cmon.md)
-
-- [CockroachDB](https://github.com/netdata/go.d.plugin/blob/master/modules/cockroachdb/integrations/cockroachdb.md)
-
-- [CouchDB](https://github.com/netdata/go.d.plugin/blob/master/modules/couchdb/integrations/couchdb.md)
-
-- [Couchbase](https://github.com/netdata/go.d.plugin/blob/master/modules/couchbase/integrations/couchbase.md)
-
-- [HANA](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/hana.md)
-
-- [Hasura GraphQL Server](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/hasura_graphql_server.md)
-
-- [InfluxDB](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/influxdb.md)
-
-- [Machbase](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/machbase.md)
-
-- [MariaDB](https://github.com/netdata/go.d.plugin/blob/master/modules/mysql/integrations/mariadb.md)
-
-- [Memcached (community)](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/memcached_community.md)
-
-- [Memcached](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/memcached/integrations/memcached.md)
-
-- [MongoDB](https://github.com/netdata/go.d.plugin/blob/master/modules/mongodb/integrations/mongodb.md)
-
-- [MySQL](https://github.com/netdata/go.d.plugin/blob/master/modules/mysql/integrations/mysql.md)
-
-- [ODBC](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/odbc.md)
-
-- [Oracle DB (community)](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/oracle_db_community.md)
-
-- [Oracle DB](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/oracledb/integrations/oracle_db.md)
-
-- [Patroni](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/patroni.md)
-
-- [Percona MySQL](https://github.com/netdata/go.d.plugin/blob/master/modules/mysql/integrations/percona_mysql.md)
-
-- [PgBouncer](https://github.com/netdata/go.d.plugin/blob/master/modules/pgbouncer/integrations/pgbouncer.md)
-
-- [Pgpool-II](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/pgpool-ii.md)
-
-- [Pika](https://github.com/netdata/go.d.plugin/blob/master/modules/pika/integrations/pika.md)
-
-- [PostgreSQL](https://github.com/netdata/go.d.plugin/blob/master/modules/postgres/integrations/postgresql.md)
-
-- [ProxySQL](https://github.com/netdata/go.d.plugin/blob/master/modules/proxysql/integrations/proxysql.md)
-
-- [Redis](https://github.com/netdata/go.d.plugin/blob/master/modules/redis/integrations/redis.md)
-
-- [RethinkDB](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/rethinkdbs/integrations/rethinkdb.md)
-
-- [RiakKV](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/riakkv/integrations/riakkv.md)
-
-- [SQL Database agnostic](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/sql_database_agnostic.md)
-
-- [Vertica](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/vertica.md)
-
-- [Warp10](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/warp10.md)
-
-- [pgBackRest](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/pgbackrest.md)
-
-### Distributed Computing Systems
-
-- [BOINC](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/boinc/integrations/boinc.md)
-
-- [Gearman](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/gearman/integrations/gearman.md)
-
-### DNS and DHCP Servers
-
-- [Akamai Edge DNS Traffic](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/akamai_edge_dns_traffic.md)
-
-- [CoreDNS](https://github.com/netdata/go.d.plugin/blob/master/modules/coredns/integrations/coredns.md)
-
-- [DNS query](https://github.com/netdata/go.d.plugin/blob/master/modules/dnsquery/integrations/dns_query.md)
-
-- [DNSBL](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/dnsbl.md)
-
-- [DNSdist](https://github.com/netdata/go.d.plugin/blob/master/modules/dnsdist/integrations/dnsdist.md)
-
-- [Dnsmasq DHCP](https://github.com/netdata/go.d.plugin/blob/master/modules/dnsmasq_dhcp/integrations/dnsmasq_dhcp.md)
-
-- [Dnsmasq](https://github.com/netdata/go.d.plugin/blob/master/modules/dnsmasq/integrations/dnsmasq.md)
-
-- [ISC Bind (RNDC)](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/bind_rndc/integrations/isc_bind_rndc.md)
-
-- [ISC DHCP](https://github.com/netdata/go.d.plugin/blob/master/modules/isc_dhcpd/integrations/isc_dhcp.md)
-
-- [Name Server Daemon](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/nsd/integrations/name_server_daemon.md)
-
-- [NextDNS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/nextdns.md)
-
-- [Pi-hole](https://github.com/netdata/go.d.plugin/blob/master/modules/pihole/integrations/pi-hole.md)
-
-- [PowerDNS Authoritative Server](https://github.com/netdata/go.d.plugin/blob/master/modules/powerdns/integrations/powerdns_authoritative_server.md)
-
-- [PowerDNS Recursor](https://github.com/netdata/go.d.plugin/blob/master/modules/powerdns_recursor/integrations/powerdns_recursor.md)
-
-- [Unbound](https://github.com/netdata/go.d.plugin/blob/master/modules/unbound/integrations/unbound.md)
-
-### eBPF
-
-- [eBPF Cachestat](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_cachestat.md)
-
-- [eBPF DCstat](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_dcstat.md)
-
-- [eBPF Disk](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_disk.md)
-
-- [eBPF Filedescriptor](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_filedescriptor.md)
-
-- [eBPF Filesystem](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_filesystem.md)
-
-- [eBPF Hardirq](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_hardirq.md)
-
-- [eBPF MDflush](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_mdflush.md)
-
-- [eBPF Mount](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_mount.md)
-
-- [eBPF OOMkill](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_oomkill.md)
-
-- [eBPF Process](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_process.md)
-
-- [eBPF Processes](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_processes.md)
-
-- [eBPF SHM](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_shm.md)
-
-- [eBPF SWAP](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_swap.md)
-
-- [eBPF Socket](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_socket.md)
-
-- [eBPF SoftIRQ](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_softirq.md)
-
-- [eBPF Sync](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_sync.md)
-
-- [eBPF VFS](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/integrations/ebpf_vfs.md)
-
-### FreeBSD
-
-- [FreeBSD NFS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/freebsd_nfs.md)
-
-- [FreeBSD RCTL-RACCT](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/freebsd_rctl-racct.md)
-
-- [dev.cpu.0.freq](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/dev.cpu.0.freq.md)
-
-- [dev.cpu.temperature](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/dev.cpu.temperature.md)
-
-- [devstat](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/devstat.md)
-
-- [getifaddrs](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/getifaddrs.md)
-
-- [getmntinfo](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/getmntinfo.md)
-
-- [hw.intrcnt](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/hw.intrcnt.md)
-
-- [ipfw](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/ipfw.md)
-
-- [kern.cp_time](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/kern.cp_time.md)
-
-- [kern.ipc.msq](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/kern.ipc.msq.md)
-
-- [kern.ipc.sem](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/kern.ipc.sem.md)
-
-- [kern.ipc.shm](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/kern.ipc.shm.md)
-
-- [net.inet.icmp.stats](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/net.inet.icmp.stats.md)
-
-- [net.inet.ip.stats](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/net.inet.ip.stats.md)
-
-- [net.inet.tcp.states](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/net.inet.tcp.states.md)
-
-- [net.inet.tcp.stats](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/net.inet.tcp.stats.md)
-
-- [net.inet.udp.stats](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/net.inet.udp.stats.md)
-
-- [net.inet6.icmp6.stats](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/net.inet6.icmp6.stats.md)
-
-- [net.inet6.ip6.stats](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/net.inet6.ip6.stats.md)
-
-- [net.isr](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/net.isr.md)
-
-- [system.ram](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/system.ram.md)
-
-- [uptime](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/uptime.md)
-
-- [vm.loadavg](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/vm.loadavg.md)
-
-- [vm.stats.sys.v_intr](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/vm.stats.sys.v_intr.md)
-
-- [vm.stats.sys.v_soft](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/vm.stats.sys.v_soft.md)
-
-- [vm.stats.sys.v_swtch](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/vm.stats.sys.v_swtch.md)
-
-- [vm.stats.vm.v_pgfaults](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/vm.stats.vm.v_pgfaults.md)
-
-- [vm.stats.vm.v_swappgs](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/vm.stats.vm.v_swappgs.md)
-
-- [vm.swap_info](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/vm.swap_info.md)
-
-- [vm.vmtotal](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/vm.vmtotal.md)
-
-- [zfs](https://github.com/netdata/netdata/blob/master/collectors/freebsd.plugin/integrations/zfs.md)
-
-### FTP Servers
-
-- [ProFTPD](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/proftpd.md)
-
-### Gaming
-
-- [BungeeCord](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/bungeecord.md)
-
-- [CS:GO](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cs:go.md)
-
-- [Minecraft](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/minecraft.md)
-
-- [OpenRCT2](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/openrct2.md)
-
-- [SpigotMC](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/spigotmc/integrations/spigotmc.md)
-
-- [Steam](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/steam.md)
-
-### Generic Data Collection
-
-- [Custom Exporter](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/custom_exporter.md)
-
-- [Excel spreadsheet](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/excel_spreadsheet.md)
-
-- [Generic Command Line Output](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/generic_command_line_output.md)
-
-- [JetBrains Floating License Server](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/jetbrains_floating_license_server.md)
-
-- [OpenWeatherMap](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/openweathermap.md)
-
-- [Pandas](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/pandas/integrations/pandas.md)
-
-- [Prometheus endpoint](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/prometheus_endpoint.md)
-
-- [SNMP devices](https://github.com/netdata/go.d.plugin/blob/master/modules/snmp/integrations/snmp_devices.md)
-
-- [Shell command](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/shell_command.md)
-
-- [Tankerkoenig API](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/tankerkoenig_api.md)
-
-- [TwinCAT ADS Web Service](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/twincat_ads_web_service.md)
-
-### Hardware Devices and Sensors
-
-- [1-Wire Sensors](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/w1sensor/integrations/1-wire_sensors.md)
-
-- [AM2320](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/am2320/integrations/am2320.md)
-
-- [AMD CPU & GPU](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/amd_cpu_&_gpu.md)
-
-- [AMD GPU](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/amd_gpu.md)
-
-- [ARM HWCPipe](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/arm_hwcpipe.md)
-
-- [CUPS](https://github.com/netdata/netdata/blob/master/collectors/cups.plugin/integrations/cups.md)
-
-- [HDD temperature](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/hddtemp/integrations/hdd_temperature.md)
-
-- [HP iLO](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/hp_ilo.md)
-
-- [IBM CryptoExpress (CEX) cards](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ibm_cryptoexpress_cex_cards.md)
-
-- [IBM Z Hardware Management Console](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ibm_z_hardware_management_console.md)
-
-- [IPMI (By SoundCloud)](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ipmi_by_soundcloud.md)
-
-- [Intelligent Platform Management Interface (IPMI)](https://github.com/netdata/netdata/blob/master/collectors/freeipmi.plugin/integrations/intelligent_platform_management_interface_ipmi.md)
-
-- [Linux Sensors (lm-sensors)](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/sensors/integrations/linux_sensors_lm-sensors.md)
-
-- [Linux Sensors (sysfs)](https://github.com/netdata/netdata/blob/master/collectors/charts.d.plugin/sensors/integrations/linux_sensors_sysfs.md)
-
-- [NVML](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/nvml.md)
-
-- [Nvidia GPU](https://github.com/netdata/go.d.plugin/blob/master/modules/nvidia_smi/integrations/nvidia_gpu.md)
-
-- [Raritan PDU](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/raritan_pdu.md)
-
-- [S.M.A.R.T.](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/smartd_log/integrations/s.m.a.r.t..md)
-
-- [ServerTech](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/servertech.md)
-
-- [Siemens S7 PLC](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/siemens_s7_plc.md)
-
-- [T-Rex NVIDIA GPU Miner](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/t-rex_nvidia_gpu_miner.md)
-
-### IoT Devices
-
-- [Airthings Waveplus air sensor](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/airthings_waveplus_air_sensor.md)
-
-- [Bobcat Miner 300](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/bobcat_miner_300.md)
-
-- [Christ Elektronik CLM5IP power panel](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/christ_elektronik_clm5ip_power_panel.md)
-
-- [CraftBeerPi](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/craftbeerpi.md)
-
-- [Dutch Electricity Smart Meter](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/dutch_electricity_smart_meter.md)
-
-- [Elgato Key Light devices.](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/elgato_key_light_devices..md)
-
-- [Energomera smart power meters](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/energomera_smart_power_meters.md)
-
-- [Helium hotspot](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/helium_hotspot.md)
-
-- [Homebridge](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/homebridge.md)
-
-- [Homey](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/homey.md)
-
-- [Jarvis Standing Desk](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/jarvis_standing_desk.md)
-
-- [MP707 USB thermometer](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/mp707_usb_thermometer.md)
-
-- [Modbus protocol](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/modbus_protocol.md)
-
-- [Monnit Sensors MQTT](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/monnit_sensors_mqtt.md)
-
-- [Nature Remo E lite devices](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/nature_remo_e_lite_devices.md)
-
-- [Netatmo sensors](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/netatmo_sensors.md)
-
-- [OpenHAB](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/openhab.md)
-
-- [Personal Weather Station](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/personal_weather_station.md)
-
-- [Philips Hue](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/philips_hue.md)
-
-- [Pimoroni Enviro+](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/pimoroni_enviro+.md)
-
-- [Powerpal devices](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/powerpal_devices.md)
-
-- [Radio Thermostat](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/radio_thermostat.md)
-
-- [SMA Inverters](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/sma_inverters.md)
-
-- [Salicru EQX inverter](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/salicru_eqx_inverter.md)
-
-- [Sense Energy](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/sense_energy.md)
-
-- [Shelly humidity sensor](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/shelly_humidity_sensor.md)
-
-- [Smart meters SML](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/smart_meters_sml.md)
-
-- [Solar logging stick](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/solar_logging_stick.md)
-
-- [SolarEdge inverters](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/solaredge_inverters.md)
-
-- [Solis Ginlong 5G inverters](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/solis_ginlong_5g_inverters.md)
-
-- [Sunspec Solar Energy](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/sunspec_solar_energy.md)
-
-- [TP-Link P110](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/tp-link_p110.md)
-
-- [Tado smart heating solution](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/tado_smart_heating_solution.md)
-
-- [Tesla Powerwall](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/tesla_powerwall.md)
-
-- [Tesla Wall Connector](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/tesla_wall_connector.md)
-
-- [Tesla vehicle](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/tesla_vehicle.md)
-
-- [Xiaomi Mi Flora](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/xiaomi_mi_flora.md)
-
-- [iqAir AirVisual air quality monitors](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/iqair_airvisual_air_quality_monitors.md)
-
-### Kubernetes
-
-- [Cilium Agent](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cilium_agent.md)
-
-- [Cilium Operator](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cilium_operator.md)
-
-- [Cilium Proxy](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cilium_proxy.md)
-
-- [Kubelet](https://github.com/netdata/go.d.plugin/blob/master/modules/k8s_kubelet/integrations/kubelet.md)
-
-- [Kubeproxy](https://github.com/netdata/go.d.plugin/blob/master/modules/k8s_kubeproxy/integrations/kubeproxy.md)
-
-- [Kubernetes Cluster Cloud Cost](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/kubernetes_cluster_cloud_cost.md)
-
-- [Kubernetes Cluster State](https://github.com/netdata/go.d.plugin/blob/master/modules/k8s_state/integrations/kubernetes_cluster_state.md)
-
-- [Kubernetes Containers](https://github.com/netdata/netdata/blob/master/collectors/cgroups.plugin/integrations/kubernetes_containers.md)
-
-- [Rancher](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/rancher.md)
-
-### Linux Systems
-
-- [CPU performance](https://github.com/netdata/netdata/blob/master/collectors/perf.plugin/integrations/cpu_performance.md)
-
-- [Disk space](https://github.com/netdata/netdata/blob/master/collectors/diskspace.plugin/integrations/disk_space.md)
-
-- [Files and directories](https://github.com/netdata/go.d.plugin/blob/master/modules/filecheck/integrations/files_and_directories.md)
-
-- [OpenRC](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/openrc.md)
-
-#### CPU
-
-- [Interrupts](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/interrupts.md)
-
-- [SoftIRQ statistics](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/softirq_statistics.md)
-
-#### Disk
-
-- [Disk Statistics](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/disk_statistics.md)
-
-- [MD RAID](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/md_raid.md)
-
-##### BTRFS
-
-- [BTRFS](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/btrfs.md)
-
-##### NFS
-
-- [NFS Client](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/nfs_client.md)
-
-- [NFS Server](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/nfs_server.md)
-
-##### ZFS
-
-- [ZFS Adaptive Replacement Cache](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/zfs_adaptive_replacement_cache.md)
-
-- [ZFS Pools](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/zfs_pools.md)
-
-#### Firewall
-
-- [Conntrack](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/conntrack.md)
-
-- [Netfilter](https://github.com/netdata/netdata/blob/master/collectors/nfacct.plugin/integrations/netfilter.md)
-
-- [Synproxy](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/synproxy.md)
-
-- [nftables](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/nftables.md)
-
-#### IPC
-
-- [Inter Process Communication](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/inter_process_communication.md)
-
-#### Kernel
-
-- [Linux kernel SLAB allocator statistics](https://github.com/netdata/netdata/blob/master/collectors/slabinfo.plugin/integrations/linux_kernel_slab_allocator_statistics.md)
-
-- [Power Capping](https://github.com/netdata/netdata/blob/master/collectors/debugfs.plugin/integrations/power_capping.md)
-
-#### Memory
-
-- [Kernel Same-Page Merging](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/kernel_same-page_merging.md)
-
-- [Linux ZSwap](https://github.com/netdata/netdata/blob/master/collectors/debugfs.plugin/integrations/linux_zswap.md)
-
-- [Memory Statistics](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/memory_statistics.md)
-
-- [Memory Usage](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/memory_usage.md)
-
-- [Memory modules (DIMMs)](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/memory_modules_dimms.md)
-
-- [Non-Uniform Memory Access](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/non-uniform_memory_access.md)
-
-- [Page types](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/page_types.md)
-
-- [System Memory Fragmentation](https://github.com/netdata/netdata/blob/master/collectors/debugfs.plugin/integrations/system_memory_fragmentation.md)
-
-- [ZRAM](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/zram.md)
-
-#### Network
-
-- [Access Points](https://github.com/netdata/netdata/blob/master/collectors/charts.d.plugin/ap/integrations/access_points.md)
-
-- [IP Virtual Server](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/ip_virtual_server.md)
-
-- [IPv6 Socket Statistics](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/ipv6_socket_statistics.md)
-
-- [InfiniBand](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/infiniband.md)
-
-- [Network interfaces](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/network_interfaces.md)
-
-- [Network statistics](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/network_statistics.md)
-
-- [SCTP Statistics](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/sctp_statistics.md)
-
-- [Socket statistics](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/socket_statistics.md)
-
-- [Softnet Statistics](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/softnet_statistics.md)
-
-- [Wireless network interfaces](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/wireless_network_interfaces.md)
-
-- [tc QoS classes](https://github.com/netdata/netdata/blob/master/collectors/tc.plugin/integrations/tc_qos_classes.md)
-
-#### Power Supply
-
-- [Power Supply](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/power_supply.md)
-
-#### Pressure
-
-- [Pressure Stall Information](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/pressure_stall_information.md)
-
-#### System
-
-- [Entropy](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/entropy.md)
-
-- [System Load Average](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/system_load_average.md)
-
-- [System Uptime](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/system_uptime.md)
-
-- [System statistics](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/integrations/system_statistics.md)
-
-### Logs Servers
-
-- [AuthLog](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/authlog.md)
-
-- [Fluentd](https://github.com/netdata/go.d.plugin/blob/master/modules/fluentd/integrations/fluentd.md)
-
-- [Graylog Server](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/graylog_server.md)
-
-- [Logstash](https://github.com/netdata/go.d.plugin/blob/master/modules/logstash/integrations/logstash.md)
-
-- [journald](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/journald.md)
-
-- [loki](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/loki.md)
-
-- [mtail](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/mtail.md)
-
-### macOS Systems
-
-- [Apple Time Machine](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/apple_time_machine.md)
-
-- [macOS](https://github.com/netdata/netdata/blob/master/collectors/macos.plugin/integrations/macos.md)
-
-### Mail Servers
-
-- [DMARC](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/dmarc.md)
-
-- [Dovecot](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/dovecot/integrations/dovecot.md)
-
-- [Exim](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/exim/integrations/exim.md)
-
-- [Halon](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/halon.md)
-
-- [Maildir](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/maildir.md)
-
-- [Postfix](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/postfix/integrations/postfix.md)
-
-### Media Services
-
-- [Discourse](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/discourse.md)
-
-- [Icecast](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/icecast/integrations/icecast.md)
-
-- [OBS Studio](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/obs_studio.md)
-
-- [RetroShare](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/retroshare/integrations/retroshare.md)
-
-- [SABnzbd](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/sabnzbd.md)
-
-- [Stream](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/stream.md)
-
-- [Twitch](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/twitch.md)
-
-- [Zulip](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/zulip.md)
-
-### Message Brokers
-
-- [ActiveMQ](https://github.com/netdata/go.d.plugin/blob/master/modules/activemq/integrations/activemq.md)
-
-- [Apache Pulsar](https://github.com/netdata/go.d.plugin/blob/master/modules/pulsar/integrations/apache_pulsar.md)
-
-- [Beanstalk](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/beanstalk/integrations/beanstalk.md)
-
-- [IBM MQ](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ibm_mq.md)
-
-- [Kafka Connect](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/kafka_connect.md)
-
-- [Kafka ZooKeeper](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/kafka_zookeeper.md)
-
-- [Kafka](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/kafka.md)
-
-- [MQTT Blackbox](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/mqtt_blackbox.md)
-
-- [RabbitMQ](https://github.com/netdata/go.d.plugin/blob/master/modules/rabbitmq/integrations/rabbitmq.md)
-
-- [Redis Queue](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/redis_queue.md)
-
-- [VerneMQ](https://github.com/netdata/go.d.plugin/blob/master/modules/vernemq/integrations/vernemq.md)
-
-- [XMPP Server](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/xmpp_server.md)
-
-- [mosquitto](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/mosquitto.md)
-
-### Networking Stack and Network Interfaces
-
-- [8430FT modem](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/8430ft_modem.md)
-
-- [A10 ACOS network devices](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/a10_acos_network_devices.md)
-
-- [Andrews & Arnold line status](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/andrews_&_arnold_line_status.md)
-
-- [Aruba devices](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/aruba_devices.md)
-
-- [Bird Routing Daemon](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/bird_routing_daemon.md)
-
-- [Checkpoint device](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/checkpoint_device.md)
-
-- [Cisco ACI](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cisco_aci.md)
-
-- [Citrix NetScaler](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/citrix_netscaler.md)
-
-- [DDWRT Routers](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ddwrt_routers.md)
-
-- [FRRouting](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/frrouting.md)
-
-- [Fortigate firewall](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/fortigate_firewall.md)
-
-- [Freifunk network](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/freifunk_network.md)
-
-- [Fritzbox network devices](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/fritzbox_network_devices.md)
-
-- [Hitron CGN series CPE](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/hitron_cgn_series_cpe.md)
-
-- [Hitron CODA Cable Modem](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/hitron_coda_cable_modem.md)
-
-- [Huawei devices](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/huawei_devices.md)
-
-- [Keepalived](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/keepalived.md)
-
-- [Meraki dashboard](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/meraki_dashboard.md)
-
-- [MikroTik devices](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/mikrotik_devices.md)
-
-- [Mikrotik RouterOS devices](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/mikrotik_routeros_devices.md)
-
-- [NetFlow](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/netflow.md)
-
-- [NetMeter](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/netmeter.md)
-
-- [Open vSwitch](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/open_vswitch.md)
-
-- [OpenROADM devices](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/openroadm_devices.md)
-
-- [RIPE Atlas](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ripe_atlas.md)
-
-- [SONiC NOS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/sonic_nos.md)
-
-- [SmartRG 808AC Cable Modem](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/smartrg_808ac_cable_modem.md)
-
-- [Starlink (SpaceX)](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/starlink_spacex.md)
-
-- [Traceroute](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/traceroute.md)
-
-- [Ubiquiti UFiber OLT](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ubiquiti_ufiber_olt.md)
-
-- [Zyxel GS1200-8](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/zyxel_gs1200-8.md)
-
-### Incident Management
-
-- [OTRS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/otrs.md)
-
-- [StatusPage](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/statuspage.md)
-
-### Observability
-
-- [Collectd](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/collectd.md)
-
-- [Dynatrace](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/dynatrace.md)
-
-- [Grafana](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/grafana.md)
-
-- [Hubble](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/hubble.md)
-
-- [Naemon](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/naemon.md)
-
-- [Nagios](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/nagios.md)
-
-- [New Relic](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/new_relic.md)
-
-### Other
-
-- [Example collector](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/example/integrations/example_collector.md)
-
-- [GitHub API rate limit](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/github_api_rate_limit.md)
-
-- [GitHub repository](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/github_repository.md)
-
-- [Netdata Agent alarms](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/alarms/integrations/netdata_agent_alarms.md)
-
-- [python.d changefinder](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/changefinder/integrations/python.d_changefinder.md)
-
-- [python.d zscores](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/zscores/integrations/python.d_zscores.md)
-
-### Processes and System Services
-
-- [Applications](https://github.com/netdata/netdata/blob/master/collectors/apps.plugin/integrations/applications.md)
-
-- [Supervisor](https://github.com/netdata/go.d.plugin/blob/master/modules/supervisord/integrations/supervisor.md)
-
-- [User Groups](https://github.com/netdata/netdata/blob/master/collectors/apps.plugin/integrations/user_groups.md)
-
-- [Users](https://github.com/netdata/netdata/blob/master/collectors/apps.plugin/integrations/users.md)
-
-### Provisioning Systems
-
-- [BOSH](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/bosh.md)
-
-- [Cloud Foundry Firehose](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cloud_foundry_firehose.md)
-
-- [Cloud Foundry](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cloud_foundry.md)
-
-- [Spacelift](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/spacelift.md)
-
-### Search Engines
-
-- [Elasticsearch](https://github.com/netdata/go.d.plugin/blob/master/modules/elasticsearch/integrations/elasticsearch.md)
-
-- [Meilisearch](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/meilisearch.md)
-
-- [OpenSearch](https://github.com/netdata/go.d.plugin/blob/master/modules/elasticsearch/integrations/opensearch.md)
-
-- [Solr](https://github.com/netdata/go.d.plugin/blob/master/modules/solr/integrations/solr.md)
-
-- [Sphinx](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/sphinx.md)
-
-### Security Systems
-
-- [Certificate Transparency](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/certificate_transparency.md)
-
-- [ClamAV daemon](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/clamav_daemon.md)
-
-- [Clamscan results](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/clamscan_results.md)
-
-- [Crowdsec](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/crowdsec.md)
-
-- [Honeypot](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/honeypot.md)
-
-- [Lynis audit reports](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/lynis_audit_reports.md)
-
-- [OpenVAS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/openvas.md)
-
-- [SSL Certificate](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ssl_certificate.md)
-
-- [Suricata](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/suricata.md)
-
-- [Vault PKI](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/vault_pki.md)
-
-### Service Discovery / Registry
-
-- [Consul](https://github.com/netdata/go.d.plugin/blob/master/modules/consul/integrations/consul.md)
-
-- [Kafka Consumer Lag](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/kafka_consumer_lag.md)
-
-- [ZooKeeper](https://github.com/netdata/go.d.plugin/blob/master/modules/zookeeper/integrations/zookeeper.md)
-
-- [etcd](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/etcd.md)
-
-### Storage, Mount Points and Filesystems
-
-- [AdaptecRAID](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/adaptec_raid/integrations/adaptecraid.md)
-
-- [Altaro Backup](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/altaro_backup.md)
-
-- [Borg backup](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/borg_backup.md)
-
-- [CVMFS clients](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cvmfs_clients.md)
-
-- [Ceph](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/ceph/integrations/ceph.md)
-
-- [Dell EMC Isilon cluster](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/dell_emc_isilon_cluster.md)
-
-- [Dell EMC ScaleIO](https://github.com/netdata/go.d.plugin/blob/master/modules/scaleio/integrations/dell_emc_scaleio.md)
-
-- [Dell EMC XtremIO cluster](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/dell_emc_xtremio_cluster.md)
-
-- [Dell PowerMax](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/dell_powermax.md)
-
-- [EOS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/eos.md)
-
-- [Generic storage enclosure tool](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/generic_storage_enclosure_tool.md)
-
-- [HDSentinel](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/hdsentinel.md)
-
-- [HP Smart Storage Arrays](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/hpssa/integrations/hp_smart_storage_arrays.md)
-
-- [Hadoop Distributed File System (HDFS)](https://github.com/netdata/go.d.plugin/blob/master/modules/hdfs/integrations/hadoop_distributed_file_system_hdfs.md)
-
-- [IBM Spectrum Virtualize](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ibm_spectrum_virtualize.md)
-
-- [IBM Spectrum](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/ibm_spectrum.md)
-
-- [IPFS](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/ipfs/integrations/ipfs.md)
-
-- [Lagerist Disk latency](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/lagerist_disk_latency.md)
-
-- [MegaCLI](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/megacli/integrations/megacli.md)
-
-- [MogileFS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/mogilefs.md)
-
-- [NVMe devices](https://github.com/netdata/go.d.plugin/blob/master/modules/nvme/integrations/nvme_devices.md)
-
-- [NetApp Solidfire](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/netapp_solidfire.md)
-
-- [Netapp ONTAP API](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/netapp_ontap_api.md)
-
-- [Samba](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/samba/integrations/samba.md)
-
-- [Starwind VSAN VSphere Edition](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/starwind_vsan_vsphere_edition.md)
-
-- [Storidge](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/storidge.md)
-
-- [Synology ActiveBackup](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/synology_activebackup.md)
-
-### Synthetic Checks
-
-- [Blackbox](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/blackbox.md)
-
-- [Domain expiration date](https://github.com/netdata/go.d.plugin/blob/master/modules/whoisquery/integrations/domain_expiration_date.md)
-
-- [HTTP Endpoints](https://github.com/netdata/go.d.plugin/blob/master/modules/httpcheck/integrations/http_endpoints.md)
-
-- [IOPing](https://github.com/netdata/netdata/blob/master/collectors/ioping.plugin/integrations/ioping.md)
-
-- [Idle OS Jitter](https://github.com/netdata/netdata/blob/master/collectors/idlejitter.plugin/integrations/idle_os_jitter.md)
-
-- [Monit](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/monit/integrations/monit.md)
-
-- [Ping](https://github.com/netdata/go.d.plugin/blob/master/modules/ping/integrations/ping.md)
-
-- [Pingdom](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/pingdom.md)
-
-- [Site 24x7](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/site_24x7.md)
-
-- [TCP Endpoints](https://github.com/netdata/go.d.plugin/blob/master/modules/portcheck/integrations/tcp_endpoints.md)
-
-- [Uptimerobot](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/uptimerobot.md)
-
-- [X.509 certificate](https://github.com/netdata/go.d.plugin/blob/master/modules/x509check/integrations/x.509_certificate.md)
-
-### System Clock and NTP
-
-- [Chrony](https://github.com/netdata/go.d.plugin/blob/master/modules/chrony/integrations/chrony.md)
-
-- [NTPd](https://github.com/netdata/go.d.plugin/blob/master/modules/ntpd/integrations/ntpd.md)
-
-- [Timex](https://github.com/netdata/netdata/blob/master/collectors/timex.plugin/integrations/timex.md)
-
-### Systemd
-
-- [Systemd Services](https://github.com/netdata/netdata/blob/master/collectors/cgroups.plugin/integrations/systemd_services.md)
-
-- [Systemd Units](https://github.com/netdata/go.d.plugin/blob/master/modules/systemdunits/integrations/systemd_units.md)
-
-- [systemd-logind users](https://github.com/netdata/go.d.plugin/blob/master/modules/logind/integrations/systemd-logind_users.md)
-
-### Task Queues
-
-- [Celery](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/celery.md)
-
-- [Mesos](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/mesos.md)
-
-- [Slurm](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/slurm.md)
-
-### Telephony Servers
-
-- [GTP](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/gtp.md)
-
-- [Kannel](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/kannel.md)
-
-- [OpenSIPS](https://github.com/netdata/netdata/blob/master/collectors/charts.d.plugin/opensips/integrations/opensips.md)
-
-### UPS
-
-- [APC UPS](https://github.com/netdata/netdata/blob/master/collectors/charts.d.plugin/apcupsd/integrations/apc_ups.md)
-
-- [Eaton UPS](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/eaton_ups.md)
-
-- [UPS (NUT)](https://github.com/netdata/go.d.plugin/blob/master/modules/upsd/integrations/ups_nut.md)
-
-### VPNs
-
-- [Fastd](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/fastd.md)
-
-- [Libreswan](https://github.com/netdata/netdata/blob/master/collectors/charts.d.plugin/libreswan/integrations/libreswan.md)
-
-- [OpenVPN status log](https://github.com/netdata/go.d.plugin/blob/master/modules/openvpn_status_log/integrations/openvpn_status_log.md)
-
-- [OpenVPN](https://github.com/netdata/go.d.plugin/blob/master/modules/openvpn/integrations/openvpn.md)
-
-- [SoftEther VPN Server](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/softether_vpn_server.md)
-
-- [Speedify CLI](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/speedify_cli.md)
-
-- [Tor](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/tor/integrations/tor.md)
-
-- [WireGuard](https://github.com/netdata/go.d.plugin/blob/master/modules/wireguard/integrations/wireguard.md)
-
-- [strongSwan](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/strongswan.md)
-
-### Web Servers and Web Proxies
-
-- [APIcast](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/apicast.md)
-
-- [Apache](https://github.com/netdata/go.d.plugin/blob/master/modules/apache/integrations/apache.md)
-
-- [Clash](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/clash.md)
-
-- [Cloudflare PCAP](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/cloudflare_pcap.md)
-
-- [Envoy](https://github.com/netdata/go.d.plugin/blob/master/modules/envoy/integrations/envoy.md)
-
-- [Gobetween](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/gobetween.md)
-
-- [HAProxy](https://github.com/netdata/go.d.plugin/blob/master/modules/haproxy/integrations/haproxy.md)
-
-- [HHVM](https://github.com/netdata/go.d.plugin/blob/master/modules/prometheus/integrations/hhvm.md)
-
-- [HTTPD](https://github.com/netdata/go.d.plugin/blob/master/modules/apache/integrations/httpd.md)
-
-- [Lighttpd](https://github.com/netdata/go.d.plugin/blob/master/modules/lighttpd/integrations/lighttpd.md)
-
-- [Litespeed](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/litespeed/integrations/litespeed.md)
-
-- [NGINX Plus](https://github.com/netdata/go.d.plugin/blob/master/modules/nginxplus/integrations/nginx_plus.md)
-
-- [NGINX VTS](https://github.com/netdata/go.d.plugin/blob/master/modules/nginxvts/integrations/nginx_vts.md)
-
-- [NGINX](https://github.com/netdata/go.d.plugin/blob/master/modules/nginx/integrations/nginx.md)
-
-- [PHP-FPM](https://github.com/netdata/go.d.plugin/blob/master/modules/phpfpm/integrations/php-fpm.md)
-
-- [Squid log files](https://github.com/netdata/go.d.plugin/blob/master/modules/squidlog/integrations/squid_log_files.md)
-
-- [Squid](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/squid/integrations/squid.md)
-
-- [Tengine](https://github.com/netdata/go.d.plugin/blob/master/modules/tengine/integrations/tengine.md)
-
-- [Tomcat](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/tomcat/integrations/tomcat.md)
-
-- [Traefik](https://github.com/netdata/go.d.plugin/blob/master/modules/traefik/integrations/traefik.md)
-
-- [Varnish](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/varnish/integrations/varnish.md)
-
-- [Web server log files](https://github.com/netdata/go.d.plugin/blob/master/modules/weblog/integrations/web_server_log_files.md)
-
-- [uWSGI](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/uwsgi/integrations/uwsgi.md)
-
-### Windows Systems
-
-- [Active Directory](https://github.com/netdata/go.d.plugin/blob/master/modules/windows/integrations/active_directory.md)
-
-- [HyperV](https://github.com/netdata/go.d.plugin/blob/master/modules/windows/integrations/hyperv.md)
-
-- [MS Exchange](https://github.com/netdata/go.d.plugin/blob/master/modules/windows/integrations/ms_exchange.md)
-
-- [MS SQL Server](https://github.com/netdata/go.d.plugin/blob/master/modules/windows/integrations/ms_sql_server.md)
-
-- [NET Framework](https://github.com/netdata/go.d.plugin/blob/master/modules/windows/integrations/net_framework.md)
-
-- [Windows](https://github.com/netdata/go.d.plugin/blob/master/modules/windows/integrations/windows.md)
diff --git a/collectors/Makefile.am b/collectors/Makefile.am
deleted file mode 100644
index 1bbb2e0ef..000000000
--- a/collectors/Makefile.am
+++ /dev/null
@@ -1,42 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-SUBDIRS = \
- plugins.d \
- apps.plugin \
- cgroups.plugin \
- charts.d.plugin \
- cups.plugin \
- debugfs.plugin \
- diskspace.plugin \
- timex.plugin \
- ioping.plugin \
- freebsd.plugin \
- freeipmi.plugin \
- idlejitter.plugin \
- log2journal \
- macos.plugin \
- nfacct.plugin \
- xenstat.plugin \
- perf.plugin \
- proc.plugin \
- python.d.plugin \
- slabinfo.plugin \
- statsd.plugin \
- ebpf.plugin \
- tc.plugin \
- systemd-journal.plugin \
- $(NULL)
-
-usercustompluginsconfigdir=$(configdir)/custom-plugins.d
-usergoconfigdir=$(configdir)/go.d
-
-# Explicitly install directories to avoid permission issues due to umask
-install-exec-local:
- $(INSTALL) -d $(DESTDIR)$(usercustompluginsconfigdir)
- $(INSTALL) -d $(DESTDIR)$(usergoconfigdir)
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/README.md b/collectors/README.md
deleted file mode 100644
index 7676ff866..000000000
--- a/collectors/README.md
+++ /dev/null
@@ -1,62 +0,0 @@
-# Collectors
-
-When Netdata starts, and with zero configuration, it auto-detects thousands of data sources and immediately collects
-per-second metrics.
-
-Netdata can immediately collect metrics from these endpoints thanks to 300+ **collectors**, which all come pre-installed
-when you [install Netdata](https://github.com/netdata/netdata/blob/master/packaging/installer/README.md).
-
-All collectors are **installed by default** with every installation of Netdata. You do not need to install
-collectors manually to collect metrics from new sources.
-See how you can [monitor anything with Netdata](https://github.com/netdata/netdata/blob/master/collectors/COLLECTORS.md).
-
-Upon startup, Netdata will **auto-detect** any application or service that has a collector, as long as both the collector
-and the app/service are configured correctly. If you don't see charts for your application, see
-our [collectors' configuration reference](https://github.com/netdata/netdata/blob/master/collectors/REFERENCE.md).
-
-## How Netdata's metrics collectors work
-
-Every collector has two primary jobs:
-
-- Look for exposed metrics at a pre- or user-defined endpoint.
-- Gather exposed metrics and use additional logic to build meaningful, interactive visualizations.
-
-If the collector finds compatible metrics exposed on the configured endpoint, it begins a per-second collection job. The
-Netdata Agent gathers these metrics, sends them to the
-[database engine for storage](https://github.com/netdata/netdata/blob/master/docs/store/change-metrics-storage.md)
-, and immediately
-[visualizes them meaningfully](https://github.com/netdata/netdata/blob/master/docs/cloud/visualize/interact-new-charts.md)
-on dashboards.
-
-Each collector comes with a pre-defined configuration that matches the default setup for that application. This endpoint
-can be a URL and port, a socket, a file, a web page, and more. The endpoint is user-configurable, as are many other
-specifics of what a given collector does.
-
-## Collector architecture and terminology
-
-- **Collectors** are the processes/programs that actually gather metrics from various sources.
-
-- **Plugins** help manage all the independent data collection processes in a variety of programming languages, based on
- their purpose and performance requirements. There are three types of plugins:
-
- - **Internal** plugins organize collectors that gather metrics from `/proc`, `/sys` and other Linux kernel sources.
- They are written in `C`, and run as threads within the Netdata daemon.
-
- - **External** plugins organize collectors that gather metrics from external processes, such as a MySQL database or
- Nginx web server. They can be written in any language, and the `netdata` daemon spawns them as long-running
- independent processes. They communicate with the daemon via pipes. All external plugins are managed by
- [plugins.d](https://github.com/netdata/netdata/blob/master/collectors/plugins.d/README.md), which provides additional management options.
-
-- **Orchestrators** are external plugins that run and manage one or more modules. They run as independent processes.
- The Go orchestrator is in active development.
-
- - [go.d.plugin](https://github.com/netdata/go.d.plugin/blob/master/README.md): An orchestrator for data
- collection modules written in `go`.
-
- - [python.d.plugin](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/README.md):
- An orchestrator for data collection modules written in `python` v2/v3.
-
- - [charts.d.plugin](https://github.com/netdata/netdata/blob/master/collectors/charts.d.plugin/README.md):
- An orchestrator for data collection modules written in`bash` v4+.
-
-- **Modules** are the individual programs controlled by an orchestrator to collect data from a specific application, or type of endpoint.
diff --git a/collectors/REFERENCE.md b/collectors/REFERENCE.md
deleted file mode 100644
index f19533f21..000000000
--- a/collectors/REFERENCE.md
+++ /dev/null
@@ -1,149 +0,0 @@
-<!--
-title: "Collectors configuration reference"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/REFERENCE.md"
-sidebar_label: "Collectors configuration"
-learn_status: "Published"
-learn_topic_type: "Tasks"
-learn_rel_path: "Configuration"
--->
-
-# Collectors configuration reference
-
-The list of supported collectors can be found in [the documentation](https://github.com/netdata/netdata/blob/master/collectors/COLLECTORS.md),
-and on [our website](https://www.netdata.cloud/integrations). The documentation of each collector provides all the
-necessary configuration options and prerequisites for that collector. In most cases, either the charts are automatically generated
-without any configuration, or you just fulfil those prerequisites and [configure the collector](#configure-a-collector).
-
-If the application you are interested in monitoring is not listed in our integrations, the collectors list includes
-the available options to
-[add your application to Netdata](https://github.com/netdata/netdata/edit/master/collectors/COLLECTORS.md#add-your-application-to-netdata).
-
-If we do support your collector but the charts described in the documentation don't appear on your dashboard, the reason will
-be one of the following:
-
-- The entire data collection plugin is disabled by default. Read how to [enable and disable plugins](#enable-and-disable-plugins)
-
-- The data collection plugin is enabled, but a specific data collection module is disabled. Read how to
- [enable and disable a specific collection module](#enable-and-disable-a-specific-collection-module).
-
-- Autodetection failed. Read how to [configure](#configure-a-collector) and [troubleshoot](#troubleshoot-a-collector) a collector.
-
-## Enable and disable plugins
-
-You can enable or disable individual plugins by opening `netdata.conf` and scrolling down to the `[plugins]` section.
-This section features a list of Netdata's plugins, with a boolean setting to enable or disable them. The exception is
-`statsd.plugin`, which has its own `[statsd]` section. Your `[plugins]` section should look similar to this:
-
-```conf
-[plugins]
- # timex = yes
- # idlejitter = yes
- # netdata monitoring = yes
- # tc = yes
- # diskspace = yes
- # proc = yes
- # cgroups = yes
- # enable running new plugins = yes
- # check for new plugins every = 60
- # slabinfo = no
- # python.d = yes
- # perf = yes
- # ioping = yes
- # fping = yes
- # nfacct = yes
- # go.d = yes
- # apps = yes
- # ebpf = yes
- # charts.d = yes
- # statsd = yes
-```
-
-By default, most plugins are enabled, so you don't need to enable them explicitly to use their collectors. To enable or
-disable any specific plugin, remove the comment (`#`) and change the boolean setting to `yes` or `no`.
-
-## Enable and disable a specific collection module
-
-You can enable/disable of the collection modules supported by `go.d`, `python.d` or `charts.d` individually, using the
-configuration file of that orchestrator. For example, you can change the behavior of the Go orchestrator, or any of its
-collectors, by editing `go.d.conf`.
-
-Use `edit-config` from your [Netdata config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory)
-to open the orchestrator primary configuration file:
-
-```bash
-cd /etc/netdata
-sudo ./edit-config go.d.conf
-```
-
-Within this file, you can either disable the orchestrator entirely (`enabled: yes`), or find a specific collector and
-enable/disable it with `yes` and `no` settings. Uncomment any line you change to ensure the Netdata daemon reads it on
-start.
-
-After you make your changes, restart the Agent with `sudo systemctl restart netdata`, or the [appropriate
-method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
-
-## Configure a collector
-
-Most collector modules come with **auto-detection**, configured to work out-of-the-box on popular operating systems with
-the default settings.
-
-However, there are cases that auto-detection fails. Usually, the reason is that the applications to be monitored do not
-allow Netdata to connect. In most of the cases, allowing the user `netdata` from `localhost` to connect and collect
-metrics, will automatically enable data collection for the application in question (it will require a Netdata restart).
-
-When Netdata starts up, each collector searches for exposed metrics on the default endpoint established by that service
-or application's standard installation procedure. For example,
-the [Nginx collector](https://github.com/netdata/go.d.plugin/blob/master/modules/nginx/README.md) searches at
-`http://127.0.0.1/stub_status` for exposed metrics in the correct format. If an Nginx web server is running and exposes
-metrics on that endpoint, the collector begins gathering them.
-
-However, not every node or infrastructure uses standard ports, paths, files, or naming conventions. You may need to
-enable or configure a collector to gather all available metrics from your systems, containers, or applications.
-
-First, [find the collector](https://github.com/netdata/netdata/blob/master/collectors/COLLECTORS.md) you want to edit
-and open its documentation. Some software has collectors written in multiple languages. In these cases, you should always
-pick the collector written in Go.
-
-Use `edit-config` from your
-[Netdata config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory)
-to open a collector's configuration file. For example, edit the Nginx collector with the following:
-
-```bash
-./edit-config go.d/nginx.conf
-```
-
-Each configuration file describes every available option and offers examples to help you tweak Netdata's settings
-according to your needs. In addition, every collector's documentation shows the exact command you need to run to
-configure that collector. Uncomment any line you change to ensure the collector's orchestrator or the Netdata daemon
-read it on start.
-
-After you make your changes, restart the Agent with `sudo systemctl restart netdata`, or the [appropriate
-method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
-
-## Troubleshoot a collector
-
-First, navigate to your plugins directory, which is usually at `/usr/libexec/netdata/plugins.d/`. If that's not the case
-on your system, open `netdata.conf` and look for the setting `plugins directory`. Once you're in the plugins directory,
-switch to the `netdata` user.
-
-```bash
-cd /usr/libexec/netdata/plugins.d/
-sudo su -s /bin/bash netdata
-```
-
-The next step is based on the collector's orchestrator.
-
-```bash
-# Go orchestrator (go.d.plugin)
-./go.d.plugin -d -m <MODULE_NAME>
-
-# Python orchestrator (python.d.plugin)
-./python.d.plugin <MODULE_NAME> debug trace
-
-# Bash orchestrator (bash.d.plugin)
-./charts.d.plugin debug 1 <MODULE_NAME>
-```
-
-The output from the relevant command will provide valuable troubleshooting information. If you can't figure out how to
-enable the collector using the details from this output, feel free to [join our Discord server](https://discord.com/invite/mPZ6WZKKG2),
-to get help from our experts.
diff --git a/collectors/apps.plugin/Makefile.am b/collectors/apps.plugin/Makefile.am
deleted file mode 100644
index 533b14dd0..000000000
--- a/collectors/apps.plugin/Makefile.am
+++ /dev/null
@@ -1,12 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
-
-dist_libconfig_DATA = \
- apps_groups.conf \
- $(NULL)
diff --git a/collectors/apps.plugin/README.md b/collectors/apps.plugin/README.md
deleted file mode 100644
index fd5371f08..000000000
--- a/collectors/apps.plugin/README.md
+++ /dev/null
@@ -1,402 +0,0 @@
-<!--
-title: "Application monitoring (apps.plugin)"
-sidebar_label: "Application monitoring "
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/apps.plugin/README.md"
-learn_status: "Published"
-learn_topic_type: "References"
-learn_rel_path: "Integrations/Monitor/System metrics"
--->
-
-# Application monitoring (apps.plugin)
-
-`apps.plugin` breaks down system resource usage to **processes**, **users** and **user groups**.
-It is enabled by default on every Netdata installation.
-
-To achieve this task, it iterates through the whole process tree, collecting resource usage information
-for every process found running.
-
-Since Netdata needs to present this information in charts and track them through time,
-instead of presenting a `top` like list, `apps.plugin` uses a pre-defined list of **process groups**
-to which it assigns all running processes. This list is customizable via `apps_groups.conf`, and Netdata
-ships with a good default for most cases (to edit it on your system run `/etc/netdata/edit-config apps_groups.conf`).
-
-So, `apps.plugin` builds a process tree (much like `ps fax` does in Linux), and groups
-processes together (evaluating both child and parent processes) so that the result is always a list with
-a predefined set of members (of course, only process groups found running are reported).
-
-> If you find that `apps.plugin` categorizes standard applications as `other`, we would be
-> glad to accept pull requests improving the defaults shipped with Netdata in `apps_groups.conf`.
-
-Unlike traditional process monitoring tools (like `top`), `apps.plugin` is able to account the resource
-utilization of exit processes. Their utilization is accounted at their currently running parents.
-So, `apps.plugin` is perfectly able to measure the resources used by shell scripts and other processes
-that fork/spawn other short-lived processes hundreds of times per second.
-
-## Charts
-
-`apps.plugin` provides charts for 3 sections:
-
-1. Per application charts as **Applications** at Netdata dashboards
-2. Per user charts as **Users** at Netdata dashboards
-3. Per user group charts as **User Groups** at Netdata dashboards
-
-Each of these sections provides the same number of charts:
-
-- CPU utilization (`apps.cpu`)
- - Total CPU usage
- - User/system CPU usage (`apps.cpu_user`/`apps.cpu_system`)
-- Disk I/O
- - Physical reads/writes (`apps.preads`/`apps.pwrites`)
- - Logical reads/writes (`apps.lreads`/`apps.lwrites`)
- - Open unique files (if a file is found open multiple times, it is counted just once, `apps.files`)
-- Memory
- - Real Memory Used (non-shared, `apps.mem`)
- - Virtual Memory Allocated (`apps.vmem`)
- - Minor page faults (i.e. memory activity, `apps.minor_faults`)
-- Processes
- - Threads running (`apps.threads`)
- - Processes running (`apps.processes`)
- - Carried over uptime (since the last Netdata Agent restart, `apps.uptime`)
- - Minimum uptime (`apps.uptime_min`)
- - Average uptime (`apps.uptime_average`)
- - Maximum uptime (`apps.uptime_max`)
- - Pipes open (`apps.pipes`)
-- Swap memory
- - Swap memory used (`apps.swap`)
- - Major page faults (i.e. swap activity, `apps.major_faults`)
-- Network
- - Sockets open (`apps.sockets`)
-
-In addition, if the [eBPF collector](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/README.md) is running, your dashboard will also show an
-additional [list of charts](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/README.md#integration-with-appsplugin) using low-level Linux
-metrics.
-
-The above are reported:
-
-- For **Applications** per target configured.
-- For **Users** per username or UID (when the username is not available).
-- For **User Groups** per group name or GID (when group name is not available).
-
-## Performance
-
-`apps.plugin` is a complex piece of software and has a lot of work to do
-We are proud that `apps.plugin` is a lot faster compared to any other similar tool,
-while collecting a lot more information for the processes, however the fact is that
-this plugin requires more CPU resources than the `netdata` daemon itself.
-
-Under Linux, for each process running, `apps.plugin` reads several `/proc` files
-per process. Doing this work per-second, especially on hosts with several thousands
-of processes, may increase the CPU resources consumed by the plugin.
-
-In such cases, you many need to lower its data collection frequency.
-
-To do this, edit `/etc/netdata/netdata.conf` and find this section:
-
-```
-[plugin:apps]
- # update every = 1
- # command options =
-```
-
-Uncomment the line `update every` and set it to a higher number. If you just set it to `2`,
-its CPU resources will be cut in half, and data collection will be once every 2 seconds.
-
-## Configuration
-
-The configuration file is `/etc/netdata/apps_groups.conf`. To edit it on your system, run `/etc/netdata/edit-config apps_groups.conf`.
-
-The configuration file works accepts multiple lines, each having this format:
-
-```txt
-group: process1 process2 ...
-```
-
-Each group can be given multiple times, to add more processes to it.
-
-For the **Applications** section, only groups configured in this file are reported.
-All other processes will be reported as `other`.
-
-For each process given, its whole process tree will be grouped, not just the process matched.
-The plugin will include both parents and children. If including the parents into the group is
-undesirable, the line `other: *` should be appended to the `apps_groups.conf`.
-
-The process names are the ones returned by:
-
-- `ps -e` or `cat /proc/PID/stat`
-- in case of substring mode (see below): `/proc/PID/cmdline`
-
-To add process names with spaces, enclose them in quotes (single or double)
-example: `'Plex Media Serv'` or `"my other process"`.
-
-You can add an asterisk `*` at the beginning and/or the end of a process:
-
-- `*name` _suffix_ mode: will search for processes ending with `name` (at `/proc/PID/stat`)
-- `name*` _prefix_ mode: will search for processes beginning with `name` (at `/proc/PID/stat`)
-- `*name*` _substring_ mode: will search for `name` in the whole command line (at `/proc/PID/cmdline`)
-
-If you enter even just one _name_ (substring), `apps.plugin` will process
-`/proc/PID/cmdline` for all processes (of course only once per process: when they are first seen).
-
-To add processes with single quotes, enclose them in double quotes: `"process with this ' single quote"`
-
-To add processes with double quotes, enclose them in single quotes: `'process with this " double quote'`
-
-If a group or process name starts with a `-`, the dimension will be hidden from the chart (cpu chart only).
-
-If a process starts with a `+`, debugging will be enabled for it (debugging produces a lot of output - do not enable it in production systems).
-
-You can add any number of groups. Only the ones found running will affect the charts generated.
-However, producing charts with hundreds of dimensions may slow down your web browser.
-
-The order of the entries in this list is important: the first that matches a process is used, so put important
-ones at the top. Processes not matched by any row, will inherit it from their parents or children.
-
-The order also controls the order of the dimensions on the generated charts (although applications started
-after apps.plugin is started, will be appended to the existing list of dimensions the `netdata` daemon maintains).
-
-There are a few command line options you can pass to `apps.plugin`. The list of available options can be acquired with the `--help` flag. The options can be set in the `netdata.conf` file. For example, to disable user and user group charts you should set
-
-```
-[plugin:apps]
- command options = without-users without-groups
-```
-
-### Integration with eBPF
-
-If you don't see charts under the **eBPF syscall** or **eBPF net** sections, you should edit your
-[`ebpf.d.conf`](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/README.md#configure-the-ebpf-collector) file to ensure the eBPF program is enabled.
-
-Also see our [guide on troubleshooting apps with eBPF
-metrics](https://github.com/netdata/netdata/blob/master/docs/guides/troubleshoot/monitor-debug-applications-ebpf.md) for ideas on how to interpret these charts in a
-few scenarios.
-
-## Permissions
-
-`apps.plugin` requires additional privileges to collect all the information it needs.
-The problem is described in issue #157.
-
-When Netdata is installed, `apps.plugin` is given the capabilities `cap_dac_read_search,cap_sys_ptrace+ep`.
-If this fails (i.e. `setcap` fails), `apps.plugin` is setuid to `root`.
-
-### linux capabilities in containers
-
-There are a few cases, like `docker` and `virtuozzo` containers, where `setcap` succeeds, but the capabilities
-are silently ignored (in `lxc` containers `setcap` fails).
-
-In this case, you will have to setuid to root `apps.plugin` by running these commands:
-
-```sh
-chown root:netdata /usr/libexec/netdata/plugins.d/apps.plugin
-chmod 4750 /usr/libexec/netdata/plugins.d/apps.plugin
-```
-
-You will have to run these, every time you update Netdata.
-
-## Security
-
-`apps.plugin` performs a hard-coded function of building the process tree in memory,
-iterating forever, collecting metrics for each running process and sending them to Netdata.
-This is a one-way communication, from `apps.plugin` to Netdata.
-
-So, since `apps.plugin` cannot be instructed by Netdata for the actions it performs,
-we think it is pretty safe to allow it to have these increased privileges.
-
-Keep in mind that `apps.plugin` will still run without escalated permissions,
-but it will not be able to collect all the information.
-
-## Application Badges
-
-You can create badges that you can embed anywhere you like, with URLs like this:
-
-```
-https://your.netdata.ip:19999/api/v1/badge.svg?chart=apps.processes&dimensions=myapp&value_color=green%3E0%7Cred
-```
-
-The color expression unescaped is this: `value_color=green>0|red`.
-
-Here is an example for the process group `sql` at `https://registry.my-netdata.io`:
-
-![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.processes&dimensions=sql&value_color=green%3E0%7Cred)
-
-Netdata is able to give you a lot more badges for your app.
-Examples below for process group `sql`:
-
-- CPU usage: ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.cpu&dimensions=sql&value_color=green=0%7Corange%3C50%7Cred)
-- Disk Physical Reads ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.preads&dimensions=sql&value_color=green%3C100%7Corange%3C1000%7Cred)
-- Disk Physical Writes ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.pwrites&dimensions=sql&value_color=green%3C100%7Corange%3C1000%7Cred)
-- Disk Logical Reads ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.lreads&dimensions=sql&value_color=green%3C100%7Corange%3C1000%7Cred)
-- Disk Logical Writes ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.lwrites&dimensions=sql&value_color=green%3C100%7Corange%3C1000%7Cred)
-- Open Files ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.fds_files&dimensions=sql&value_color=green%3E30%7Cred)
-- Real Memory ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.mem&dimensions=sql&value_color=green%3C100%7Corange%3C200%7Cred)
-- Virtual Memory ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.vmem&dimensions=sql&value_color=green%3C100%7Corange%3C1000%7Cred)
-- Swap Memory ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.swap&dimensions=sql&value_color=green=0%7Cred)
-- Minor Page Faults ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.minor_faults&dimensions=sql&value_color=green%3C100%7Corange%3C1000%7Cred)
-- Processes ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.processes&dimensions=sql&value_color=green%3E0%7Cred)
-- Threads ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.threads&dimensions=sql&value_color=green%3E=28%7Cred)
-- Major Faults (swap activity) ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.major_faults&dimensions=sql&value_color=green=0%7Cred)
-- Open Pipes ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.fds_pipes&dimensions=sql&value_color=green=0%7Cred)
-- Open Sockets ![image](https://registry.my-netdata.io/api/v1/badge.svg?chart=apps.fds_sockets&dimensions=sql&value_color=green%3E=3%7Cred)
-
-For more information about badges check [Generating Badges](https://github.com/netdata/netdata/blob/master/web/api/badges/README.md)
-
-## Comparison with console tools
-
-SSH to a server running Netdata and execute this:
-
-```sh
-while true; do ls -l /var/run >/dev/null; done
-```
-
-In most systems `/var/run` is a `tmpfs` device, so there is nothing that can stop this command
-from consuming entirely one of the CPU cores of the machine.
-
-As we will see below, **none** of the console performance monitoring tools can report that this
-command is using 100% CPU. They do report of course that the CPU is busy, but **they fail to
-identify the process that consumes so much CPU**.
-
-Here is what common Linux console monitoring tools report:
-
-### top
-
-`top` reports that `bash` is using just 14%.
-
-If you check the total system CPU utilization, it says there is no idle CPU at all, but `top`
-fails to provide a breakdown of the CPU consumption in the system. The sum of the CPU utilization
-of all processes reported by `top`, is 15.6%.
-
-```
-top - 18:46:28 up 3 days, 20:14, 2 users, load average: 0.22, 0.05, 0.02
-Tasks: 76 total, 2 running, 74 sleeping, 0 stopped, 0 zombie
-%Cpu(s): 32.8 us, 65.6 sy, 0.0 ni, 0.0 id, 0.0 wa, 1.3 hi, 0.3 si, 0.0 st
-KiB Mem : 1016576 total, 244112 free, 52012 used, 720452 buff/cache
-KiB Swap: 0 total, 0 free, 0 used. 753712 avail Mem
-
- PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
-12789 root 20 0 14980 4180 3020 S 14.0 0.4 0:02.82 bash
- 9 root 20 0 0 0 0 S 1.0 0.0 0:22.36 rcuos/0
- 642 netdata 20 0 132024 20112 2660 S 0.3 2.0 14:26.29 netdata
-12522 netdata 20 0 9508 2476 1828 S 0.3 0.2 0:02.26 apps.plugin
- 1 root 20 0 67196 10216 7500 S 0.0 1.0 0:04.83 systemd
- 2 root 20 0 0 0 0 S 0.0 0.0 0:00.00 kthreadd
-```
-
-### htop
-
-Exactly like `top`, `htop` is providing an incomplete breakdown of the system CPU utilization.
-
-```
- CPU[||||||||||||||||||||||||100.0%] Tasks: 27, 11 thr; 2 running
- Mem[||||||||||||||||||||85.4M/993M] Load average: 1.16 0.88 0.90
- Swp[ 0K/0K] Uptime: 3 days, 21:37:03
-
- PID USER PRI NI VIRT RES SHR S CPU% MEM% TIME+ Command
-12789 root 20 0 15104 4484 3208 S 14.0 0.4 10:57.15 -bash
- 7024 netdata 20 0 9544 2480 1744 S 0.7 0.2 0:00.88 /usr/libexec/netd
- 7009 netdata 20 0 138M 21016 2712 S 0.7 2.1 0:00.89 /usr/sbin/netdata
- 7012 netdata 20 0 138M 21016 2712 S 0.0 2.1 0:00.31 /usr/sbin/netdata
- 563 root 20 0 308M 202M 202M S 0.0 20.4 1:00.81 /usr/lib/systemd/
- 7019 netdata 20 0 138M 21016 2712 S 0.0 2.1 0:00.14 /usr/sbin/netdata
-```
-
-### atop
-
-`atop` also fails to break down CPU usage.
-
-```
-ATOP - localhost 2016/12/10 20:11:27 ----------- 10s elapsed
-PRC | sys 1.13s | user 0.43s | #proc 75 | #zombie 0 | #exit 5383 |
-CPU | sys 67% | user 31% | irq 2% | idle 0% | wait 0% |
-CPL | avg1 1.34 | avg5 1.05 | avg15 0.96 | csw 51346 | intr 10508 |
-MEM | tot 992.8M | free 211.5M | cache 470.0M | buff 87.2M | slab 164.7M |
-SWP | tot 0.0M | free 0.0M | | vmcom 207.6M | vmlim 496.4M |
-DSK | vda | busy 0% | read 0 | write 4 | avio 1.50 ms |
-NET | transport | tcpi 16 | tcpo 15 | udpi 0 | udpo 0 |
-NET | network | ipi 16 | ipo 15 | ipfrw 0 | deliv 16 |
-NET | eth0 ---- | pcki 16 | pcko 15 | si 1 Kbps | so 4 Kbps |
-
- PID SYSCPU USRCPU VGROW RGROW RDDSK WRDSK ST EXC S CPU CMD 1/600
-12789 0.98s 0.40s 0K 0K 0K 336K -- - S 14% bash
- 9 0.08s 0.00s 0K 0K 0K 0K -- - S 1% rcuos/0
- 7024 0.03s 0.00s 0K 0K 0K 0K -- - S 0% apps.plugin
- 7009 0.01s 0.01s 0K 0K 0K 4K -- - S 0% netdata
-```
-
-### glances
-
-And the same is true for `glances`. The system runs at 100%, but `glances` reports only 17%
-per process utilization.
-
-Note also, that being a `python` program, `glances` uses 1.6% CPU while it runs.
-
-```
-localhost Uptime: 3 days, 21:42:00
-
-CPU [100.0%] CPU 100.0% MEM 23.7% SWAP 0.0% LOAD 1-core
-MEM [ 23.7%] user: 30.9% total: 993M total: 0 1 min: 1.18
-SWAP [ 0.0%] system: 67.8% used: 236M used: 0 5 min: 1.08
- idle: 0.0% free: 757M free: 0 15 min: 1.00
-
-NETWORK Rx/s Tx/s TASKS 75 (90 thr), 1 run, 74 slp, 0 oth
-eth0 168b 2Kb
-eth1 0b 0b CPU% MEM% PID USER NI S Command
-lo 0b 0b 13.5 0.4 12789 root 0 S -bash
- 1.6 2.2 7025 root 0 R /usr/bin/python /u
-DISK I/O R/s W/s 1.0 0.0 9 root 0 S rcuos/0
-vda1 0 4K 0.3 0.2 7024 netdata 0 S /usr/libexec/netda
- 0.3 0.0 7 root 0 S rcu_sched
-FILE SYS Used Total 0.3 2.1 7009 netdata 0 S /usr/sbin/netdata
-/ (vda1) 1.56G 29.5G 0.0 0.0 17 root 0 S oom_reaper
-```
-
-### why does this happen?
-
-All the console tools report usage based on the processes found running *at the moment they
-examine the process tree*. So, they see just one `ls` command, which is actually very quick
-with minor CPU utilization. But the shell, is spawning hundreds of them, one after another
-(much like shell scripts do).
-
-### What does Netdata report?
-
-The total CPU utilization of the system:
-
-![image](https://cloud.githubusercontent.com/assets/2662304/21076212/9198e5a6-bf2e-11e6-9bc0-6bdea25befb2.png)
-<br/>***Figure 1**: The system overview section at Netdata, just a few seconds after the command was run*
-
-And at the applications `apps.plugin` breaks down CPU usage per application:
-
-![image](https://cloud.githubusercontent.com/assets/2662304/21076220/c9687848-bf2e-11e6-8d81-348592c5aca2.png)
-<br/>***Figure 2**: The Applications section at Netdata, just a few seconds after the command was run*
-
-So, the `ssh` session is using 95% CPU time.
-
-Why `ssh`?
-
-`apps.plugin` groups all processes based on its configuration file.
-The default configuration has nothing for `bash`, but it has for `sshd`, so Netdata accumulates
-all ssh sessions to a dimension on the charts, called `ssh`. This includes all the processes in
-the process tree of `sshd`, **including the exited children**.
-
-> Distributions based on `systemd`, provide another way to get cpu utilization per user session
-> or service running: control groups, or cgroups, commonly used as part of containers
-> `apps.plugin` does not use these mechanisms. The process grouping made by `apps.plugin` works
-> on any Linux, `systemd` based or not.
-
-#### a more technical description of how Netdata works
-
-Netdata reads `/proc/<pid>/stat` for all processes, once per second and extracts `utime` and
-`stime` (user and system cpu utilization), much like all the console tools do.
-
-But it also extracts `cutime` and `cstime` that account the user and system time of the exit children of each process.
-By keeping a map in memory of the whole process tree, it is capable of assigning the right time to every process, taking
-into account all its exited children.
-
-It is tricky, since a process may be running for 1 hour and once it exits, its parent should not
-receive the whole 1 hour of cpu time in just 1 second - you have to subtract the cpu time that has
-been reported for it prior to this iteration.
-
-It is even trickier, because walking through the entire process tree takes some time itself. So,
-if you sum the CPU utilization of all processes, you might have more CPU time than the reported
-total cpu time of the system. Netdata solves this, by adapting the per process cpu utilization to
-the total of the system. [Netdata adds charts that document this normalization](https://london.my-netdata.io/default.html#menu_netdata_submenu_apps_plugin).
-
-
diff --git a/collectors/apps.plugin/apps_plugin.c b/collectors/apps.plugin/apps_plugin.c
deleted file mode 100644
index 5bcda84f4..000000000
--- a/collectors/apps.plugin/apps_plugin.c
+++ /dev/null
@@ -1,5422 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-/*
- * netdata apps.plugin
- * (C) Copyright 2023 Netdata Inc.
- * Released under GPL v3+
- */
-
-#include "collectors/all.h"
-#include "libnetdata/libnetdata.h"
-#include "libnetdata/required_dummies.h"
-
-#define APPS_PLUGIN_PROCESSES_FUNCTION_DESCRIPTION "Detailed information on the currently running processes."
-
-#define APPS_PLUGIN_FUNCTIONS() do { \
- fprintf(stdout, PLUGINSD_KEYWORD_FUNCTION " \"processes\" %d \"%s\"\n", PLUGINS_FUNCTIONS_TIMEOUT_DEFAULT, APPS_PLUGIN_PROCESSES_FUNCTION_DESCRIPTION); \
- } while(0)
-
-#define APPS_PLUGIN_GLOBAL_FUNCTIONS() do { \
- fprintf(stdout, PLUGINSD_KEYWORD_FUNCTION " GLOBAL \"processes\" %d \"%s\"\n", PLUGINS_FUNCTIONS_TIMEOUT_DEFAULT, APPS_PLUGIN_PROCESSES_FUNCTION_DESCRIPTION); \
- } while(0)
-
-// ----------------------------------------------------------------------------
-// debugging
-
-static int debug_enabled = 0;
-static inline void debug_log_int(const char *fmt, ... ) {
- va_list args;
-
- fprintf( stderr, "apps.plugin: ");
- va_start( args, fmt );
- vfprintf( stderr, fmt, args );
- va_end( args );
-
- fputc('\n', stderr);
-}
-
-#ifdef NETDATA_INTERNAL_CHECKS
-
-#define debug_log(fmt, args...) do { if(unlikely(debug_enabled)) debug_log_int(fmt, ##args); } while(0)
-
-#else
-
-static inline void debug_log_dummy(void) {}
-#define debug_log(fmt, args...) debug_log_dummy()
-
-#endif
-
-
-// ----------------------------------------------------------------------------
-
-#ifdef __FreeBSD__
-#include <sys/user.h>
-#endif
-
-// ----------------------------------------------------------------------------
-// per O/S configuration
-
-// the minimum PID of the system
-// this is also the pid of the init process
-#define INIT_PID 1
-
-// if the way apps.plugin will work, will read the entire process list,
-// including the resource utilization of each process, instantly
-// set this to 1
-// when set to 0, apps.plugin builds a sort list of processes, in order
-// to process children processes, before parent processes
-#ifdef __FreeBSD__
-#define ALL_PIDS_ARE_READ_INSTANTLY 1
-#else
-#define ALL_PIDS_ARE_READ_INSTANTLY 0
-#endif
-
-// ----------------------------------------------------------------------------
-// string lengths
-
-#define MAX_COMPARE_NAME 100
-#define MAX_NAME 100
-#define MAX_CMDLINE 16384
-
-// ----------------------------------------------------------------------------
-// the rates we are going to send to netdata will have this detail a value of:
-// - 1 will send just integer parts to netdata
-// - 100 will send 2 decimal points
-// - 1000 will send 3 decimal points
-// etc.
-#define RATES_DETAIL 10000ULL
-
-// ----------------------------------------------------------------------------
-// factor for calculating correct CPU time values depending on units of raw data
-static unsigned int time_factor = 0;
-
-// ----------------------------------------------------------------------------
-// to avoid reallocating too frequently, we can increase the number of spare
-// file descriptors used by processes.
-// IMPORTANT:
-// having a lot of spares, increases the CPU utilization of the plugin.
-#define MAX_SPARE_FDS 1
-
-// ----------------------------------------------------------------------------
-// command line options
-
-static int
- update_every = 1,
- enable_guest_charts = 0,
-#ifdef __FreeBSD__
- enable_file_charts = 0,
-#else
- enable_file_charts = 1,
- max_fds_cache_seconds = 60,
-#endif
- enable_function_cmdline = 0,
- enable_detailed_uptime_charts = 0,
- enable_users_charts = 1,
- enable_groups_charts = 1,
- include_exited_childs = 1;
-
-// will be changed to getenv(NETDATA_USER_CONFIG_DIR) if it exists
-static char *user_config_dir = CONFIG_DIR;
-static char *stock_config_dir = LIBCONFIG_DIR;
-
-// some variables for keeping track of processes count by states
-typedef enum {
- PROC_STATUS_RUNNING = 0,
- PROC_STATUS_SLEEPING_D, // uninterruptible sleep
- PROC_STATUS_SLEEPING, // interruptible sleep
- PROC_STATUS_ZOMBIE,
- PROC_STATUS_STOPPED,
- PROC_STATUS_END, //place holder for ending enum fields
-} proc_state;
-
-#ifndef __FreeBSD__
-static proc_state proc_state_count[PROC_STATUS_END];
-static const char *proc_states[] = {
- [PROC_STATUS_RUNNING] = "running",
- [PROC_STATUS_SLEEPING] = "sleeping_interruptible",
- [PROC_STATUS_SLEEPING_D] = "sleeping_uninterruptible",
- [PROC_STATUS_ZOMBIE] = "zombie",
- [PROC_STATUS_STOPPED] = "stopped",
- };
-#endif
-
-// ----------------------------------------------------------------------------
-// internal flags
-// handled in code (automatically set)
-
-// log each problem once per process
-// log flood protection flags (log_thrown)
-typedef enum __attribute__((packed)) {
- PID_LOG_IO = (1 << 0),
- PID_LOG_STATUS = (1 << 1),
- PID_LOG_CMDLINE = (1 << 2),
- PID_LOG_FDS = (1 << 3),
- PID_LOG_STAT = (1 << 4),
- PID_LOG_LIMITS = (1 << 5),
- PID_LOG_LIMITS_DETAIL = (1 << 6),
-} PID_LOG;
-
-static int
- show_guest_time = 0, // 1 when guest values are collected
- show_guest_time_old = 0,
- proc_pid_cmdline_is_needed = 0; // 1 when we need to read /proc/cmdline
-
-
-// ----------------------------------------------------------------------------
-// internal counters
-
-static size_t
- global_iterations_counter = 1,
- calls_counter = 0,
- file_counter = 0,
- filenames_allocated_counter = 0,
- inodes_changed_counter = 0,
- links_changed_counter = 0,
- targets_assignment_counter = 0;
-
-
-// ----------------------------------------------------------------------------
-// Normalization
-//
-// With normalization we lower the collected metrics by a factor to make them
-// match the total utilization of the system.
-// The discrepancy exists because apps.plugin needs some time to collect all
-// the metrics. This results in utilization that exceeds the total utilization
-// of the system.
-//
-// During normalization, we align the per-process utilization, to the total of
-// the system. We first consume the exited children utilization and it the
-// collected values is above the total, we proportionally scale each reported
-// metric.
-
-// the total system time, as reported by /proc/stat
-static kernel_uint_t
- global_utime = 0,
- global_stime = 0,
- global_gtime = 0;
-
-// the normalization ratios, as calculated by normalize_utilization()
-NETDATA_DOUBLE
- utime_fix_ratio = 1.0,
- stime_fix_ratio = 1.0,
- gtime_fix_ratio = 1.0,
- minflt_fix_ratio = 1.0,
- majflt_fix_ratio = 1.0,
- cutime_fix_ratio = 1.0,
- cstime_fix_ratio = 1.0,
- cgtime_fix_ratio = 1.0,
- cminflt_fix_ratio = 1.0,
- cmajflt_fix_ratio = 1.0;
-
-
-struct pid_on_target {
- int32_t pid;
- struct pid_on_target *next;
-};
-
-struct openfds {
- kernel_uint_t files;
- kernel_uint_t pipes;
- kernel_uint_t sockets;
- kernel_uint_t inotifies;
- kernel_uint_t eventfds;
- kernel_uint_t timerfds;
- kernel_uint_t signalfds;
- kernel_uint_t eventpolls;
- kernel_uint_t other;
-};
-
-#define pid_openfds_sum(p) ((p)->openfds.files + (p)->openfds.pipes + (p)->openfds.sockets + (p)->openfds.inotifies + (p)->openfds.eventfds + (p)->openfds.timerfds + (p)->openfds.signalfds + (p)->openfds.eventpolls + (p)->openfds.other)
-
-struct pid_limits {
-// kernel_uint_t max_cpu_time;
-// kernel_uint_t max_file_size;
-// kernel_uint_t max_data_size;
-// kernel_uint_t max_stack_size;
-// kernel_uint_t max_core_file_size;
-// kernel_uint_t max_resident_set;
-// kernel_uint_t max_processes;
- kernel_uint_t max_open_files;
-// kernel_uint_t max_locked_memory;
-// kernel_uint_t max_address_space;
-// kernel_uint_t max_file_locks;
-// kernel_uint_t max_pending_signals;
-// kernel_uint_t max_msgqueue_size;
-// kernel_uint_t max_nice_priority;
-// kernel_uint_t max_realtime_priority;
-// kernel_uint_t max_realtime_timeout;
-};
-
-// ----------------------------------------------------------------------------
-// target
-//
-// target is the structure that processes are aggregated to be reported
-// to netdata.
-//
-// - Each entry in /etc/apps_groups.conf creates a target.
-// - Each user and group used by a process in the system, creates a target.
-
-struct target {
- char compare[MAX_COMPARE_NAME + 1];
- uint32_t comparehash;
- size_t comparelen;
-
- char id[MAX_NAME + 1];
- uint32_t idhash;
-
- char name[MAX_NAME + 1];
- char clean_name[MAX_NAME + 1]; // sanitized name used in chart id (need to replace at least dots)
- uid_t uid;
- gid_t gid;
-
- bool is_other;
-
- kernel_uint_t minflt;
- kernel_uint_t cminflt;
- kernel_uint_t majflt;
- kernel_uint_t cmajflt;
- kernel_uint_t utime;
- kernel_uint_t stime;
- kernel_uint_t gtime;
- kernel_uint_t cutime;
- kernel_uint_t cstime;
- kernel_uint_t cgtime;
- kernel_uint_t num_threads;
- // kernel_uint_t rss;
-
- kernel_uint_t status_vmsize;
- kernel_uint_t status_vmrss;
- kernel_uint_t status_vmshared;
- kernel_uint_t status_rssfile;
- kernel_uint_t status_rssshmem;
- kernel_uint_t status_vmswap;
- kernel_uint_t status_voluntary_ctxt_switches;
- kernel_uint_t status_nonvoluntary_ctxt_switches;
-
- kernel_uint_t io_logical_bytes_read;
- kernel_uint_t io_logical_bytes_written;
- kernel_uint_t io_read_calls;
- kernel_uint_t io_write_calls;
- kernel_uint_t io_storage_bytes_read;
- kernel_uint_t io_storage_bytes_written;
- kernel_uint_t io_cancelled_write_bytes;
-
- int *target_fds;
- int target_fds_size;
-
- struct openfds openfds;
-
- NETDATA_DOUBLE max_open_files_percent;
-
- kernel_uint_t starttime;
- kernel_uint_t collected_starttime;
- kernel_uint_t uptime_min;
- kernel_uint_t uptime_sum;
- kernel_uint_t uptime_max;
-
- unsigned int processes; // how many processes have been merged to this
- int exposed; // if set, we have sent this to netdata
- int hidden; // if set, we set the hidden flag on the dimension
- int debug_enabled;
- int ends_with;
- int starts_with; // if set, the compare string matches only the
- // beginning of the command
-
- struct pid_on_target *root_pid; // list of aggregated pids for target debugging
-
- struct target *target; // the one that will be reported to netdata
- struct target *next;
-};
-
-struct target
- *apps_groups_default_target = NULL, // the default target
- *apps_groups_root_target = NULL, // apps_groups.conf defined
- *users_root_target = NULL, // users
- *groups_root_target = NULL; // user groups
-
-size_t
- apps_groups_targets_count = 0; // # of apps_groups.conf targets
-
-
-// ----------------------------------------------------------------------------
-// pid_stat
-//
-// structure to store data for each process running
-// see: man proc for the description of the fields
-
-struct pid_fd {
- int fd;
-
-#ifndef __FreeBSD__
- ino_t inode;
- char *filename;
- uint32_t link_hash;
- size_t cache_iterations_counter;
- size_t cache_iterations_reset;
-#endif
-};
-
-struct pid_stat {
- int32_t pid;
- int32_t ppid;
- // int32_t pgrp;
- // int32_t session;
- // int32_t tty_nr;
- // int32_t tpgid;
- // uint64_t flags;
-
- char state;
-
- char comm[MAX_COMPARE_NAME + 1];
- char *cmdline;
-
- // these are raw values collected
- kernel_uint_t minflt_raw;
- kernel_uint_t cminflt_raw;
- kernel_uint_t majflt_raw;
- kernel_uint_t cmajflt_raw;
- kernel_uint_t utime_raw;
- kernel_uint_t stime_raw;
- kernel_uint_t gtime_raw; // guest_time
- kernel_uint_t cutime_raw;
- kernel_uint_t cstime_raw;
- kernel_uint_t cgtime_raw; // cguest_time
-
- // these are rates
- kernel_uint_t minflt;
- kernel_uint_t cminflt;
- kernel_uint_t majflt;
- kernel_uint_t cmajflt;
- kernel_uint_t utime;
- kernel_uint_t stime;
- kernel_uint_t gtime;
- kernel_uint_t cutime;
- kernel_uint_t cstime;
- kernel_uint_t cgtime;
-
- // int64_t priority;
- // int64_t nice;
- int32_t num_threads;
- // int64_t itrealvalue;
- kernel_uint_t collected_starttime;
- // kernel_uint_t vsize;
- // kernel_uint_t rss;
- // kernel_uint_t rsslim;
- // kernel_uint_t starcode;
- // kernel_uint_t endcode;
- // kernel_uint_t startstack;
- // kernel_uint_t kstkesp;
- // kernel_uint_t kstkeip;
- // uint64_t signal;
- // uint64_t blocked;
- // uint64_t sigignore;
- // uint64_t sigcatch;
- // uint64_t wchan;
- // uint64_t nswap;
- // uint64_t cnswap;
- // int32_t exit_signal;
- // int32_t processor;
- // uint32_t rt_priority;
- // uint32_t policy;
- // kernel_uint_t delayacct_blkio_ticks;
-
- uid_t uid;
- gid_t gid;
-
- kernel_uint_t status_voluntary_ctxt_switches_raw;
- kernel_uint_t status_nonvoluntary_ctxt_switches_raw;
-
- kernel_uint_t status_vmsize;
- kernel_uint_t status_vmrss;
- kernel_uint_t status_vmshared;
- kernel_uint_t status_rssfile;
- kernel_uint_t status_rssshmem;
- kernel_uint_t status_vmswap;
- kernel_uint_t status_voluntary_ctxt_switches;
- kernel_uint_t status_nonvoluntary_ctxt_switches;
-#ifndef __FreeBSD__
- ARL_BASE *status_arl;
-#endif
-
- kernel_uint_t io_logical_bytes_read_raw;
- kernel_uint_t io_logical_bytes_written_raw;
- kernel_uint_t io_read_calls_raw;
- kernel_uint_t io_write_calls_raw;
- kernel_uint_t io_storage_bytes_read_raw;
- kernel_uint_t io_storage_bytes_written_raw;
- kernel_uint_t io_cancelled_write_bytes_raw;
-
- kernel_uint_t io_logical_bytes_read;
- kernel_uint_t io_logical_bytes_written;
- kernel_uint_t io_read_calls;
- kernel_uint_t io_write_calls;
- kernel_uint_t io_storage_bytes_read;
- kernel_uint_t io_storage_bytes_written;
- kernel_uint_t io_cancelled_write_bytes;
-
- kernel_uint_t uptime;
-
- struct pid_fd *fds; // array of fds it uses
- size_t fds_size; // the size of the fds array
-
- struct openfds openfds;
- struct pid_limits limits;
-
- NETDATA_DOUBLE openfds_limits_percent;
-
- int sortlist; // higher numbers = top on the process tree
- // each process gets a unique number
-
- int children_count; // number of processes directly referencing this
- int keeploops; // increases by 1 every time keep is 1 and updated 0
-
- PID_LOG log_thrown;
-
- bool keep; // true when we need to keep this process in memory even after it exited
- bool updated; // true when the process is currently running
- bool merged; // true when it has been merged to its parent
- bool read; // true when we have already read this process for this iteration
- bool matched_by_config;
-
- struct target *target; // app_groups.conf targets
- struct target *user_target; // uid based targets
- struct target *group_target; // gid based targets
-
- usec_t stat_collected_usec;
- usec_t last_stat_collected_usec;
-
- usec_t io_collected_usec;
- usec_t last_io_collected_usec;
- usec_t last_limits_collected_usec;
-
- char *fds_dirname; // the full directory name in /proc/PID/fd
-
- char *stat_filename;
- char *status_filename;
- char *io_filename;
- char *cmdline_filename;
- char *limits_filename;
-
- struct pid_stat *parent;
- struct pid_stat *prev;
- struct pid_stat *next;
-};
-
-size_t pagesize;
-
-kernel_uint_t global_uptime;
-
-static struct pid_stat
- *root_of_pids = NULL, // global list of all processes running
- **all_pids = NULL; // to avoid allocations, we pre-allocate
- // a pointer for each pid in the entire pid space.
-
-static size_t
- all_pids_count = 0; // the number of processes running
-
-#if (ALL_PIDS_ARE_READ_INSTANTLY == 0)
-// Another pre-allocated list of all possible pids.
-// We need it to pids and assign them a unique sortlist id, so that we
-// read parents before children. This is needed to prevent a situation where
-// a child is found running, but until we read its parent, it has exited and
-// its parent has accumulated its resources.
-static pid_t
- *all_pids_sortlist = NULL;
-#endif
-
-// ----------------------------------------------------------------------------
-// file descriptor
-//
-// this is used to keep a global list of all open files of the system.
-// it is needed in order to calculate the unique files processes have open.
-
-#define FILE_DESCRIPTORS_INCREASE_STEP 100
-
-// types for struct file_descriptor->type
-typedef enum fd_filetype {
- FILETYPE_OTHER,
- FILETYPE_FILE,
- FILETYPE_PIPE,
- FILETYPE_SOCKET,
- FILETYPE_INOTIFY,
- FILETYPE_EVENTFD,
- FILETYPE_EVENTPOLL,
- FILETYPE_TIMERFD,
- FILETYPE_SIGNALFD
-} FD_FILETYPE;
-
-struct file_descriptor {
- avl_t avl;
-
-#ifdef NETDATA_INTERNAL_CHECKS
- uint32_t magic;
-#endif /* NETDATA_INTERNAL_CHECKS */
-
- const char *name;
- uint32_t hash;
-
- FD_FILETYPE type;
- int count;
- int pos;
-} *all_files = NULL;
-
-static int
- all_files_len = 0,
- all_files_size = 0;
-
-// ----------------------------------------------------------------------------
-// read users and groups from files
-
-struct user_or_group_id {
- avl_t avl;
-
- union {
- uid_t uid;
- gid_t gid;
- } id;
-
- char *name;
-
- int updated;
-
- struct user_or_group_id * next;
-};
-
-enum user_or_group_id_type {
- USER_ID,
- GROUP_ID
-};
-
-struct user_or_group_ids{
- enum user_or_group_id_type type;
-
- avl_tree_type index;
- struct user_or_group_id *root;
-
- char filename[FILENAME_MAX + 1];
-};
-
-int user_id_compare(void* a, void* b) {
- if(((struct user_or_group_id *)a)->id.uid < ((struct user_or_group_id *)b)->id.uid)
- return -1;
-
- else if(((struct user_or_group_id *)a)->id.uid > ((struct user_or_group_id *)b)->id.uid)
- return 1;
-
- else
- return 0;
-}
-
-struct user_or_group_ids all_user_ids = {
- .type = USER_ID,
-
- .index = {
- NULL,
- user_id_compare
- },
-
- .root = NULL,
-
- .filename = "",
-};
-
-int group_id_compare(void* a, void* b) {
- if(((struct user_or_group_id *)a)->id.gid < ((struct user_or_group_id *)b)->id.gid)
- return -1;
-
- else if(((struct user_or_group_id *)a)->id.gid > ((struct user_or_group_id *)b)->id.gid)
- return 1;
-
- else
- return 0;
-}
-
-struct user_or_group_ids all_group_ids = {
- .type = GROUP_ID,
-
- .index = {
- NULL,
- group_id_compare
- },
-
- .root = NULL,
-
- .filename = "",
-};
-
-int file_changed(const struct stat *statbuf, struct timespec *last_modification_time) {
- if(likely(statbuf->st_mtim.tv_sec == last_modification_time->tv_sec &&
- statbuf->st_mtim.tv_nsec == last_modification_time->tv_nsec)) return 0;
-
- last_modification_time->tv_sec = statbuf->st_mtim.tv_sec;
- last_modification_time->tv_nsec = statbuf->st_mtim.tv_nsec;
-
- return 1;
-}
-
-int read_user_or_group_ids(struct user_or_group_ids *ids, struct timespec *last_modification_time) {
- struct stat statbuf;
- if(unlikely(stat(ids->filename, &statbuf)))
- return 1;
- else
- if(likely(!file_changed(&statbuf, last_modification_time))) return 0;
-
- procfile *ff = procfile_open(ids->filename, " :\t", PROCFILE_FLAG_DEFAULT);
- if(unlikely(!ff)) return 1;
-
- ff = procfile_readall(ff);
- if(unlikely(!ff)) return 1;
-
- size_t line, lines = procfile_lines(ff);
-
- for(line = 0; line < lines ;line++) {
- size_t words = procfile_linewords(ff, line);
- if(unlikely(words < 3)) continue;
-
- char *name = procfile_lineword(ff, line, 0);
- if(unlikely(!name || !*name)) continue;
-
- char *id_string = procfile_lineword(ff, line, 2);
- if(unlikely(!id_string || !*id_string)) continue;
-
-
- struct user_or_group_id *user_or_group_id = callocz(1, sizeof(struct user_or_group_id));
-
- if(ids->type == USER_ID)
- user_or_group_id->id.uid = (uid_t) str2ull(id_string, NULL);
- else
- user_or_group_id->id.gid = (uid_t) str2ull(id_string, NULL);
-
- user_or_group_id->name = strdupz(name);
- user_or_group_id->updated = 1;
-
- struct user_or_group_id *existing_user_id = NULL;
-
- if(likely(ids->root))
- existing_user_id = (struct user_or_group_id *)avl_search(&ids->index, (avl_t *) user_or_group_id);
-
- if(unlikely(existing_user_id)) {
- freez(existing_user_id->name);
- existing_user_id->name = user_or_group_id->name;
- existing_user_id->updated = 1;
- freez(user_or_group_id);
- }
- else {
- if(unlikely(avl_insert(&ids->index, (avl_t *) user_or_group_id) != (void *) user_or_group_id)) {
- netdata_log_error("INTERNAL ERROR: duplicate indexing of id during realloc");
- };
-
- user_or_group_id->next = ids->root;
- ids->root = user_or_group_id;
- }
- }
-
- procfile_close(ff);
-
- // remove unused ids
- struct user_or_group_id *user_or_group_id = ids->root, *prev_user_id = NULL;
-
- while(user_or_group_id) {
- if(unlikely(!user_or_group_id->updated)) {
- if(unlikely((struct user_or_group_id *)avl_remove(&ids->index, (avl_t *) user_or_group_id) != user_or_group_id))
- netdata_log_error("INTERNAL ERROR: removal of unused id from index, removed a different id");
-
- if(prev_user_id)
- prev_user_id->next = user_or_group_id->next;
- else
- ids->root = user_or_group_id->next;
-
- freez(user_or_group_id->name);
- freez(user_or_group_id);
-
- if(prev_user_id)
- user_or_group_id = prev_user_id->next;
- else
- user_or_group_id = ids->root;
- }
- else {
- user_or_group_id->updated = 0;
-
- prev_user_id = user_or_group_id;
- user_or_group_id = user_or_group_id->next;
- }
- }
-
- return 0;
-}
-
-// ----------------------------------------------------------------------------
-// apps_groups.conf
-// aggregate all processes in groups, to have a limited number of dimensions
-
-static struct target *get_users_target(uid_t uid) {
- struct target *w;
- for(w = users_root_target ; w ; w = w->next)
- if(w->uid == uid) return w;
-
- w = callocz(sizeof(struct target), 1);
- snprintfz(w->compare, MAX_COMPARE_NAME, "%u", uid);
- w->comparehash = simple_hash(w->compare);
- w->comparelen = strlen(w->compare);
-
- snprintfz(w->id, MAX_NAME, "%u", uid);
- w->idhash = simple_hash(w->id);
-
- struct user_or_group_id user_id_to_find, *user_or_group_id = NULL;
- user_id_to_find.id.uid = uid;
-
- if(*netdata_configured_host_prefix) {
- static struct timespec last_passwd_modification_time;
- int ret = read_user_or_group_ids(&all_user_ids, &last_passwd_modification_time);
-
- if(likely(!ret && all_user_ids.index.root))
- user_or_group_id = (struct user_or_group_id *)avl_search(&all_user_ids.index, (avl_t *) &user_id_to_find);
- }
-
- if(user_or_group_id && user_or_group_id->name && *user_or_group_id->name) {
- snprintfz(w->name, MAX_NAME, "%s", user_or_group_id->name);
- }
- else {
- struct passwd *pw = getpwuid(uid);
- if(!pw || !pw->pw_name || !*pw->pw_name)
- snprintfz(w->name, MAX_NAME, "%u", uid);
- else
- snprintfz(w->name, MAX_NAME, "%s", pw->pw_name);
- }
-
- strncpyz(w->clean_name, w->name, MAX_NAME);
- netdata_fix_chart_name(w->clean_name);
-
- w->uid = uid;
-
- w->next = users_root_target;
- users_root_target = w;
-
- debug_log("added uid %u ('%s') target", w->uid, w->name);
-
- return w;
-}
-
-struct target *get_groups_target(gid_t gid)
-{
- struct target *w;
- for(w = groups_root_target ; w ; w = w->next)
- if(w->gid == gid) return w;
-
- w = callocz(sizeof(struct target), 1);
- snprintfz(w->compare, MAX_COMPARE_NAME, "%u", gid);
- w->comparehash = simple_hash(w->compare);
- w->comparelen = strlen(w->compare);
-
- snprintfz(w->id, MAX_NAME, "%u", gid);
- w->idhash = simple_hash(w->id);
-
- struct user_or_group_id group_id_to_find, *group_id = NULL;
- group_id_to_find.id.gid = gid;
-
- if(*netdata_configured_host_prefix) {
- static struct timespec last_group_modification_time;
- int ret = read_user_or_group_ids(&all_group_ids, &last_group_modification_time);
-
- if(likely(!ret && all_group_ids.index.root))
- group_id = (struct user_or_group_id *)avl_search(&all_group_ids.index, (avl_t *) &group_id_to_find);
- }
-
- if(group_id && group_id->name && *group_id->name) {
- snprintfz(w->name, MAX_NAME, "%s", group_id->name);
- }
- else {
- struct group *gr = getgrgid(gid);
- if(!gr || !gr->gr_name || !*gr->gr_name)
- snprintfz(w->name, MAX_NAME, "%u", gid);
- else
- snprintfz(w->name, MAX_NAME, "%s", gr->gr_name);
- }
-
- strncpyz(w->clean_name, w->name, MAX_NAME);
- netdata_fix_chart_name(w->clean_name);
-
- w->gid = gid;
-
- w->next = groups_root_target;
- groups_root_target = w;
-
- debug_log("added gid %u ('%s') target", w->gid, w->name);
-
- return w;
-}
-
-// find or create a new target
-// there are targets that are just aggregated to other target (the second argument)
-static struct target *get_apps_groups_target(const char *id, struct target *target, const char *name) {
- int tdebug = 0, thidden = target?target->hidden:0, ends_with = 0;
- const char *nid = id;
-
- // extract the options
- while(nid[0] == '-' || nid[0] == '+' || nid[0] == '*') {
- if(nid[0] == '-') thidden = 1;
- if(nid[0] == '+') tdebug = 1;
- if(nid[0] == '*') ends_with = 1;
- nid++;
- }
- uint32_t hash = simple_hash(id);
-
- // find if it already exists
- struct target *w, *last = apps_groups_root_target;
- for(w = apps_groups_root_target ; w ; w = w->next) {
- if(w->idhash == hash && strncmp(nid, w->id, MAX_NAME) == 0)
- return w;
-
- last = w;
- }
-
- // find an existing target
- if(unlikely(!target)) {
- while(*name == '-') {
- if(*name == '-') thidden = 1;
- name++;
- }
-
- for(target = apps_groups_root_target ; target != NULL ; target = target->next) {
- if(!target->target && strcmp(name, target->name) == 0)
- break;
- }
-
- if(unlikely(debug_enabled)) {
- if(unlikely(target))
- debug_log("REUSING TARGET NAME '%s' on ID '%s'", target->name, target->id);
- else
- debug_log("NEW TARGET NAME '%s' on ID '%s'", name, id);
- }
- }
-
- if(target && target->target)
- fatal("Internal Error: request to link process '%s' to target '%s' which is linked to target '%s'", id, target->id, target->target->id);
-
- w = callocz(sizeof(struct target), 1);
- strncpyz(w->id, nid, MAX_NAME);
- w->idhash = simple_hash(w->id);
-
- if(unlikely(!target))
- // copy the name
- strncpyz(w->name, name, MAX_NAME);
- else
- // copy the id
- strncpyz(w->name, nid, MAX_NAME);
-
- // dots are used to distinguish chart type and id in streaming, so we should replace them
- strncpyz(w->clean_name, w->name, MAX_NAME);
- netdata_fix_chart_name(w->clean_name);
- for (char *d = w->clean_name; *d; d++) {
- if (*d == '.')
- *d = '_';
- }
-
- strncpyz(w->compare, nid, MAX_COMPARE_NAME);
- size_t len = strlen(w->compare);
- if(w->compare[len - 1] == '*') {
- w->compare[len - 1] = '\0';
- w->starts_with = 1;
- }
- w->ends_with = ends_with;
-
- if(w->starts_with && w->ends_with)
- proc_pid_cmdline_is_needed = 1;
-
- w->comparehash = simple_hash(w->compare);
- w->comparelen = strlen(w->compare);
-
- w->hidden = thidden;
-#ifdef NETDATA_INTERNAL_CHECKS
- w->debug_enabled = tdebug;
-#else
- if(tdebug)
- fprintf(stderr, "apps.plugin has been compiled without debugging\n");
-#endif
- w->target = target;
-
- // append it, to maintain the order in apps_groups.conf
- if(last) last->next = w;
- else apps_groups_root_target = w;
-
- debug_log("ADDING TARGET ID '%s', process name '%s' (%s), aggregated on target '%s', options: %s %s"
- , w->id
- , w->compare, (w->starts_with && w->ends_with)?"substring":((w->starts_with)?"prefix":((w->ends_with)?"suffix":"exact"))
- , w->target?w->target->name:w->name
- , (w->hidden)?"hidden":"-"
- , (w->debug_enabled)?"debug":"-"
- );
-
- return w;
-}
-
-// read the apps_groups.conf file
-static int read_apps_groups_conf(const char *path, const char *file)
-{
- char filename[FILENAME_MAX + 1];
-
- snprintfz(filename, FILENAME_MAX, "%s/apps_%s.conf", path, file);
-
- debug_log("process groups file: '%s'", filename);
-
- // ----------------------------------------
-
- procfile *ff = procfile_open(filename, " :\t", PROCFILE_FLAG_DEFAULT);
- if(!ff) return 1;
-
- procfile_set_quotes(ff, "'\"");
-
- ff = procfile_readall(ff);
- if(!ff)
- return 1;
-
- size_t line, lines = procfile_lines(ff);
-
- for(line = 0; line < lines ;line++) {
- size_t word, words = procfile_linewords(ff, line);
- if(!words) continue;
-
- char *name = procfile_lineword(ff, line, 0);
- if(!name || !*name) continue;
-
- // find a possibly existing target
- struct target *w = NULL;
-
- // loop through all words, skipping the first one (the name)
- for(word = 0; word < words ;word++) {
- char *s = procfile_lineword(ff, line, word);
- if(!s || !*s) continue;
- if(*s == '#') break;
-
- // is this the first word? skip it
- if(s == name) continue;
-
- // add this target
- struct target *n = get_apps_groups_target(s, w, name);
- if(!n) {
- netdata_log_error("Cannot create target '%s' (line %zu, word %zu)", s, line, word);
- continue;
- }
-
- // just some optimization
- // to avoid searching for a target for each process
- if(!w) w = n->target?n->target:n;
- }
- }
-
- procfile_close(ff);
-
- apps_groups_default_target = get_apps_groups_target("p+!o@w#e$i^r&7*5(-i)l-o_", NULL, "other"); // match nothing
- if(!apps_groups_default_target)
- fatal("Cannot create default target");
- apps_groups_default_target->is_other = true;
-
- // allow the user to override group 'other'
- if(apps_groups_default_target->target)
- apps_groups_default_target = apps_groups_default_target->target;
-
- return 0;
-}
-
-
-// ----------------------------------------------------------------------------
-// struct pid_stat management
-static inline void init_pid_fds(struct pid_stat *p, size_t first, size_t size);
-
-static inline struct pid_stat *get_pid_entry(pid_t pid) {
- if(unlikely(all_pids[pid]))
- return all_pids[pid];
-
- struct pid_stat *p = callocz(sizeof(struct pid_stat), 1);
- p->fds = mallocz(sizeof(struct pid_fd) * MAX_SPARE_FDS);
- p->fds_size = MAX_SPARE_FDS;
- init_pid_fds(p, 0, p->fds_size);
- p->pid = pid;
-
- DOUBLE_LINKED_LIST_APPEND_ITEM_UNSAFE(root_of_pids, p, prev, next);
-
- all_pids[pid] = p;
- all_pids_count++;
-
- return p;
-}
-
-static inline void del_pid_entry(pid_t pid) {
- struct pid_stat *p = all_pids[pid];
-
- if(unlikely(!p)) {
- netdata_log_error("attempted to free pid %d that is not allocated.", pid);
- return;
- }
-
- debug_log("process %d %s exited, deleting it.", pid, p->comm);
-
- DOUBLE_LINKED_LIST_REMOVE_ITEM_UNSAFE(root_of_pids, p, prev, next);
-
- // free the filename
-#ifndef __FreeBSD__
- {
- size_t i;
- for(i = 0; i < p->fds_size; i++)
- if(p->fds[i].filename)
- freez(p->fds[i].filename);
- }
-#endif
- freez(p->fds);
-
- freez(p->fds_dirname);
- freez(p->stat_filename);
- freez(p->status_filename);
- freez(p->limits_filename);
-#ifndef __FreeBSD__
- arl_free(p->status_arl);
-#endif
- freez(p->io_filename);
- freez(p->cmdline_filename);
- freez(p->cmdline);
- freez(p);
-
- all_pids[pid] = NULL;
- all_pids_count--;
-}
-
-// ----------------------------------------------------------------------------
-
-static inline int managed_log(struct pid_stat *p, PID_LOG log, int status) {
- if(unlikely(!status)) {
- // netdata_log_error("command failed log %u, errno %d", log, errno);
-
- if(unlikely(debug_enabled || errno != ENOENT)) {
- if(unlikely(debug_enabled || !(p->log_thrown & log))) {
- p->log_thrown |= log;
- switch(log) {
- case PID_LOG_IO:
- #ifdef __FreeBSD__
- netdata_log_error("Cannot fetch process %d I/O info (command '%s')", p->pid, p->comm);
- #else
- netdata_log_error("Cannot process %s/proc/%d/io (command '%s')", netdata_configured_host_prefix, p->pid, p->comm);
- #endif
- break;
-
- case PID_LOG_STATUS:
- #ifdef __FreeBSD__
- netdata_log_error("Cannot fetch process %d status info (command '%s')", p->pid, p->comm);
- #else
- netdata_log_error("Cannot process %s/proc/%d/status (command '%s')", netdata_configured_host_prefix, p->pid, p->comm);
- #endif
- break;
-
- case PID_LOG_CMDLINE:
- #ifdef __FreeBSD__
- netdata_log_error("Cannot fetch process %d command line (command '%s')", p->pid, p->comm);
- #else
- netdata_log_error("Cannot process %s/proc/%d/cmdline (command '%s')", netdata_configured_host_prefix, p->pid, p->comm);
- #endif
- break;
-
- case PID_LOG_FDS:
- #ifdef __FreeBSD__
- netdata_log_error("Cannot fetch process %d files (command '%s')", p->pid, p->comm);
- #else
- netdata_log_error("Cannot process entries in %s/proc/%d/fd (command '%s')", netdata_configured_host_prefix, p->pid, p->comm);
- #endif
- break;
-
- case PID_LOG_LIMITS:
- #ifdef __FreeBSD__
- ;
- #else
- netdata_log_error("Cannot process %s/proc/%d/limits (command '%s')", netdata_configured_host_prefix, p->pid, p->comm);
- #endif
-
- case PID_LOG_STAT:
- break;
-
- default:
- netdata_log_error("unhandled error for pid %d, command '%s'", p->pid, p->comm);
- break;
- }
- }
- }
- errno = 0;
- }
- else if(unlikely(p->log_thrown & log)) {
- // netdata_log_error("unsetting log %u on pid %d", log, p->pid);
- p->log_thrown &= ~log;
- }
-
- return status;
-}
-
-static inline void assign_target_to_pid(struct pid_stat *p) {
- targets_assignment_counter++;
-
- uint32_t hash = simple_hash(p->comm);
- size_t pclen = strlen(p->comm);
-
- struct target *w;
- for(w = apps_groups_root_target; w ; w = w->next) {
- // if(debug_enabled || (p->target && p->target->debug_enabled)) debug_log_int("\t\tcomparing '%s' with '%s'", w->compare, p->comm);
-
- // find it - 4 cases:
- // 1. the target is not a pattern
- // 2. the target has the prefix
- // 3. the target has the suffix
- // 4. the target is something inside cmdline
-
- if(unlikely(( (!w->starts_with && !w->ends_with && w->comparehash == hash && !strcmp(w->compare, p->comm))
- || (w->starts_with && !w->ends_with && !strncmp(w->compare, p->comm, w->comparelen))
- || (!w->starts_with && w->ends_with && pclen >= w->comparelen && !strcmp(w->compare, &p->comm[pclen - w->comparelen]))
- || (proc_pid_cmdline_is_needed && w->starts_with && w->ends_with && p->cmdline && strstr(p->cmdline, w->compare))
- ))) {
-
- p->matched_by_config = true;
- if(w->target) p->target = w->target;
- else p->target = w;
-
- if(debug_enabled || (p->target && p->target->debug_enabled))
- debug_log_int("%s linked to target %s", p->comm, p->target->name);
-
- break;
- }
- }
-}
-
-
-// ----------------------------------------------------------------------------
-// update pids from proc
-
-static inline int read_proc_pid_cmdline(struct pid_stat *p) {
- static char cmdline[MAX_CMDLINE + 1];
-
-#ifdef __FreeBSD__
- size_t i, bytes = MAX_CMDLINE;
- int mib[4];
-
- mib[0] = CTL_KERN;
- mib[1] = KERN_PROC;
- mib[2] = KERN_PROC_ARGS;
- mib[3] = p->pid;
- if (unlikely(sysctl(mib, 4, cmdline, &bytes, NULL, 0)))
- goto cleanup;
-#else
- if(unlikely(!p->cmdline_filename)) {
- char filename[FILENAME_MAX + 1];
- snprintfz(filename, FILENAME_MAX, "%s/proc/%d/cmdline", netdata_configured_host_prefix, p->pid);
- p->cmdline_filename = strdupz(filename);
- }
-
- int fd = open(p->cmdline_filename, procfile_open_flags, 0666);
- if(unlikely(fd == -1)) goto cleanup;
-
- ssize_t i, bytes = read(fd, cmdline, MAX_CMDLINE);
- close(fd);
-
- if(unlikely(bytes < 0)) goto cleanup;
-#endif
-
- cmdline[bytes] = '\0';
- for(i = 0; i < bytes ; i++) {
- if(unlikely(!cmdline[i])) cmdline[i] = ' ';
- }
-
- if(p->cmdline) freez(p->cmdline);
- p->cmdline = strdupz(cmdline);
-
- debug_log("Read file '%s' contents: %s", p->cmdline_filename, p->cmdline);
-
- return 1;
-
-cleanup:
- // copy the command to the command line
- if(p->cmdline) freez(p->cmdline);
- p->cmdline = strdupz(p->comm);
- return 0;
-}
-
-// ----------------------------------------------------------------------------
-// macro to calculate the incremental rate of a value
-// each parameter is accessed only ONCE - so it is safe to pass function calls
-// or other macros as parameters
-
-#define incremental_rate(rate_variable, last_kernel_variable, new_kernel_value, collected_usec, last_collected_usec) { \
- kernel_uint_t _new_tmp = new_kernel_value; \
- (rate_variable) = (_new_tmp - (last_kernel_variable)) * (USEC_PER_SEC * RATES_DETAIL) / ((collected_usec) - (last_collected_usec)); \
- (last_kernel_variable) = _new_tmp; \
- }
-
-// the same macro for struct pid members
-#define pid_incremental_rate(type, var, value) \
- incremental_rate(var, var##_raw, value, p->type##_collected_usec, p->last_##type##_collected_usec)
-
-
-// ----------------------------------------------------------------------------
-
-#ifndef __FreeBSD__
-struct arl_callback_ptr {
- struct pid_stat *p;
- procfile *ff;
- size_t line;
-};
-
-void arl_callback_status_uid(const char *name, uint32_t hash, const char *value, void *dst) {
- (void)name; (void)hash; (void)value;
- struct arl_callback_ptr *aptr = (struct arl_callback_ptr *)dst;
- if(unlikely(procfile_linewords(aptr->ff, aptr->line) < 5)) return;
-
- //const char *real_uid = procfile_lineword(aptr->ff, aptr->line, 1);
- const char *effective_uid = procfile_lineword(aptr->ff, aptr->line, 2);
- //const char *saved_uid = procfile_lineword(aptr->ff, aptr->line, 3);
- //const char *filesystem_uid = procfile_lineword(aptr->ff, aptr->line, 4);
-
- if(likely(effective_uid && *effective_uid))
- aptr->p->uid = (uid_t)str2l(effective_uid);
-}
-
-void arl_callback_status_gid(const char *name, uint32_t hash, const char *value, void *dst) {
- (void)name; (void)hash; (void)value;
- struct arl_callback_ptr *aptr = (struct arl_callback_ptr *)dst;
- if(unlikely(procfile_linewords(aptr->ff, aptr->line) < 5)) return;
-
- //const char *real_gid = procfile_lineword(aptr->ff, aptr->line, 1);
- const char *effective_gid = procfile_lineword(aptr->ff, aptr->line, 2);
- //const char *saved_gid = procfile_lineword(aptr->ff, aptr->line, 3);
- //const char *filesystem_gid = procfile_lineword(aptr->ff, aptr->line, 4);
-
- if(likely(effective_gid && *effective_gid))
- aptr->p->gid = (uid_t)str2l(effective_gid);
-}
-
-void arl_callback_status_vmsize(const char *name, uint32_t hash, const char *value, void *dst) {
- (void)name; (void)hash; (void)value;
- struct arl_callback_ptr *aptr = (struct arl_callback_ptr *)dst;
- if(unlikely(procfile_linewords(aptr->ff, aptr->line) < 3)) return;
-
- aptr->p->status_vmsize = str2kernel_uint_t(procfile_lineword(aptr->ff, aptr->line, 1));
-}
-
-void arl_callback_status_vmswap(const char *name, uint32_t hash, const char *value, void *dst) {
- (void)name; (void)hash; (void)value;
- struct arl_callback_ptr *aptr = (struct arl_callback_ptr *)dst;
- if(unlikely(procfile_linewords(aptr->ff, aptr->line) < 3)) return;
-
- aptr->p->status_vmswap = str2kernel_uint_t(procfile_lineword(aptr->ff, aptr->line, 1));
-}
-
-void arl_callback_status_vmrss(const char *name, uint32_t hash, const char *value, void *dst) {
- (void)name; (void)hash; (void)value;
- struct arl_callback_ptr *aptr = (struct arl_callback_ptr *)dst;
- if(unlikely(procfile_linewords(aptr->ff, aptr->line) < 3)) return;
-
- aptr->p->status_vmrss = str2kernel_uint_t(procfile_lineword(aptr->ff, aptr->line, 1));
-}
-
-void arl_callback_status_rssfile(const char *name, uint32_t hash, const char *value, void *dst) {
- (void)name; (void)hash; (void)value;
- struct arl_callback_ptr *aptr = (struct arl_callback_ptr *)dst;
- if(unlikely(procfile_linewords(aptr->ff, aptr->line) < 3)) return;
-
- aptr->p->status_rssfile = str2kernel_uint_t(procfile_lineword(aptr->ff, aptr->line, 1));
-}
-
-void arl_callback_status_rssshmem(const char *name, uint32_t hash, const char *value, void *dst) {
- (void)name; (void)hash; (void)value;
- struct arl_callback_ptr *aptr = (struct arl_callback_ptr *)dst;
- if(unlikely(procfile_linewords(aptr->ff, aptr->line) < 3)) return;
-
- aptr->p->status_rssshmem = str2kernel_uint_t(procfile_lineword(aptr->ff, aptr->line, 1));
-}
-
-void arl_callback_status_voluntary_ctxt_switches(const char *name, uint32_t hash, const char *value, void *dst) {
- (void)name; (void)hash; (void)value;
- struct arl_callback_ptr *aptr = (struct arl_callback_ptr *)dst;
- if(unlikely(procfile_linewords(aptr->ff, aptr->line) < 2)) return;
-
- struct pid_stat *p = aptr->p;
- pid_incremental_rate(stat, p->status_voluntary_ctxt_switches, str2kernel_uint_t(procfile_lineword(aptr->ff, aptr->line, 1)));
-}
-
-void arl_callback_status_nonvoluntary_ctxt_switches(const char *name, uint32_t hash, const char *value, void *dst) {
- (void)name; (void)hash; (void)value;
- struct arl_callback_ptr *aptr = (struct arl_callback_ptr *)dst;
- if(unlikely(procfile_linewords(aptr->ff, aptr->line) < 2)) return;
-
- struct pid_stat *p = aptr->p;
- pid_incremental_rate(stat, p->status_nonvoluntary_ctxt_switches, str2kernel_uint_t(procfile_lineword(aptr->ff, aptr->line, 1)));
-}
-
-static void update_proc_state_count(char proc_state) {
- switch (proc_state) {
- case 'S':
- proc_state_count[PROC_STATUS_SLEEPING] += 1;
- break;
- case 'R':
- proc_state_count[PROC_STATUS_RUNNING] += 1;
- break;
- case 'D':
- proc_state_count[PROC_STATUS_SLEEPING_D] += 1;
- break;
- case 'Z':
- proc_state_count[PROC_STATUS_ZOMBIE] += 1;
- break;
- case 'T':
- proc_state_count[PROC_STATUS_STOPPED] += 1;
- break;
- default:
- break;
- }
-}
-#endif // !__FreeBSD__
-
-#define MAX_PROC_PID_LIMITS 8192
-#define PROC_PID_LIMITS_MAX_OPEN_FILES_KEY "\nMax open files "
-
-static inline kernel_uint_t get_proc_pid_limits_limit(char *buf, const char *key, size_t key_len, kernel_uint_t def) {
- char *line = strstr(buf, key);
- if(!line)
- return def;
-
- char *v = &line[key_len];
- while(isspace(*v)) v++;
-
- if(strcmp(v, "unlimited") == 0)
- return 0;
-
- return str2ull(v, NULL);
-}
-
-static inline int read_proc_pid_limits(struct pid_stat *p, void *ptr) {
- (void)ptr;
-
-#ifdef __FreeBSD__
- return 0;
-#else
- static char proc_pid_limits_buffer[MAX_PROC_PID_LIMITS + 1];
- int ret = 0;
- bool read_limits = false;
-
- errno = 0;
- proc_pid_limits_buffer[0] = '\0';
-
- kernel_uint_t all_fds = pid_openfds_sum(p);
- if(all_fds < p->limits.max_open_files / 2 && p->io_collected_usec > p->last_limits_collected_usec && p->io_collected_usec - p->last_limits_collected_usec <= 60 * USEC_PER_SEC) {
- // too frequent, we want to collect limits once per minute
- ret = 1;
- goto cleanup;
- }
-
- if(unlikely(!p->limits_filename)) {
- char filename[FILENAME_MAX + 1];
- snprintfz(filename, FILENAME_MAX, "%s/proc/%d/limits", netdata_configured_host_prefix, p->pid);
- p->limits_filename = strdupz(filename);
- }
-
- int fd = open(p->limits_filename, procfile_open_flags, 0666);
- if(unlikely(fd == -1)) goto cleanup;
-
- ssize_t bytes = read(fd, proc_pid_limits_buffer, MAX_PROC_PID_LIMITS);
- close(fd);
-
- if(bytes <= 0)
- goto cleanup;
-
- // make it '\0' terminated
- if(bytes < MAX_PROC_PID_LIMITS)
- proc_pid_limits_buffer[bytes] = '\0';
- else
- proc_pid_limits_buffer[MAX_PROC_PID_LIMITS - 1] = '\0';
-
- p->limits.max_open_files = get_proc_pid_limits_limit(proc_pid_limits_buffer, PROC_PID_LIMITS_MAX_OPEN_FILES_KEY, sizeof(PROC_PID_LIMITS_MAX_OPEN_FILES_KEY) - 1, 0);
- if(p->limits.max_open_files == 1) {
- // it seems a bug in the kernel or something similar
- // it sets max open files to 1 but the number of files
- // the process has open are more than 1...
- // https://github.com/netdata/netdata/issues/15443
- p->limits.max_open_files = 0;
- ret = 1;
- goto cleanup;
- }
-
- p->last_limits_collected_usec = p->io_collected_usec;
- read_limits = true;
-
- ret = 1;
-
-cleanup:
- if(p->limits.max_open_files)
- p->openfds_limits_percent = (NETDATA_DOUBLE)all_fds * 100.0 / (NETDATA_DOUBLE)p->limits.max_open_files;
- else
- p->openfds_limits_percent = 0.0;
-
- if(p->openfds_limits_percent > 100.0) {
- if(!(p->log_thrown & PID_LOG_LIMITS_DETAIL)) {
- char *line;
-
- if(!read_limits) {
- proc_pid_limits_buffer[0] = '\0';
- line = "NOT READ";
- }
- else {
- line = strstr(proc_pid_limits_buffer, PROC_PID_LIMITS_MAX_OPEN_FILES_KEY);
- if (line) {
- line++; // skip the initial newline
-
- char *end = strchr(line, '\n');
- if (end)
- *end = '\0';
- }
- }
-
- netdata_log_info(
- "FDS_LIMITS: PID %d (%s) is using "
- "%0.2f %% of its fds limits, "
- "open fds = %"PRIu64 "("
- "files = %"PRIu64 ", "
- "pipes = %"PRIu64 ", "
- "sockets = %"PRIu64", "
- "inotifies = %"PRIu64", "
- "eventfds = %"PRIu64", "
- "timerfds = %"PRIu64", "
- "signalfds = %"PRIu64", "
- "eventpolls = %"PRIu64" "
- "other = %"PRIu64" "
- "), open fds limit = %"PRIu64", "
- "%s, "
- "original line [%s]",
- p->pid, p->comm, p->openfds_limits_percent, all_fds,
- p->openfds.files,
- p->openfds.pipes,
- p->openfds.sockets,
- p->openfds.inotifies,
- p->openfds.eventfds,
- p->openfds.timerfds,
- p->openfds.signalfds,
- p->openfds.eventpolls,
- p->openfds.other,
- p->limits.max_open_files,
- read_limits ? "and we have read the limits AFTER counting the fds"
- : "but we have read the limits BEFORE counting the fds",
- line);
-
- p->log_thrown |= PID_LOG_LIMITS_DETAIL;
- }
- }
- else
- p->log_thrown &= ~PID_LOG_LIMITS_DETAIL;
-
- return ret;
-#endif
-}
-
-static inline int read_proc_pid_status(struct pid_stat *p, void *ptr) {
- p->status_vmsize = 0;
- p->status_vmrss = 0;
- p->status_vmshared = 0;
- p->status_rssfile = 0;
- p->status_rssshmem = 0;
- p->status_vmswap = 0;
- p->status_voluntary_ctxt_switches = 0;
- p->status_nonvoluntary_ctxt_switches = 0;
-
-#ifdef __FreeBSD__
- struct kinfo_proc *proc_info = (struct kinfo_proc *)ptr;
-
- p->uid = proc_info->ki_uid;
- p->gid = proc_info->ki_groups[0];
- p->status_vmsize = proc_info->ki_size / 1024; // in KiB
- p->status_vmrss = proc_info->ki_rssize * pagesize / 1024; // in KiB
- // TODO: what about shared and swap memory on FreeBSD?
- return 1;
-#else
- (void)ptr;
-
- static struct arl_callback_ptr arl_ptr;
- static procfile *ff = NULL;
-
- if(unlikely(!p->status_arl)) {
- p->status_arl = arl_create("/proc/pid/status", NULL, 60);
- arl_expect_custom(p->status_arl, "Uid", arl_callback_status_uid, &arl_ptr);
- arl_expect_custom(p->status_arl, "Gid", arl_callback_status_gid, &arl_ptr);
- arl_expect_custom(p->status_arl, "VmSize", arl_callback_status_vmsize, &arl_ptr);
- arl_expect_custom(p->status_arl, "VmRSS", arl_callback_status_vmrss, &arl_ptr);
- arl_expect_custom(p->status_arl, "RssFile", arl_callback_status_rssfile, &arl_ptr);
- arl_expect_custom(p->status_arl, "RssShmem", arl_callback_status_rssshmem, &arl_ptr);
- arl_expect_custom(p->status_arl, "VmSwap", arl_callback_status_vmswap, &arl_ptr);
- arl_expect_custom(p->status_arl, "voluntary_ctxt_switches", arl_callback_status_voluntary_ctxt_switches, &arl_ptr);
- arl_expect_custom(p->status_arl, "nonvoluntary_ctxt_switches", arl_callback_status_nonvoluntary_ctxt_switches, &arl_ptr);
- }
-
-
- if(unlikely(!p->status_filename)) {
- char filename[FILENAME_MAX + 1];
- snprintfz(filename, FILENAME_MAX, "%s/proc/%d/status", netdata_configured_host_prefix, p->pid);
- p->status_filename = strdupz(filename);
- }
-
- ff = procfile_reopen(ff, p->status_filename, (!ff)?" \t:,-()/":NULL, PROCFILE_FLAG_NO_ERROR_ON_FILE_IO);
- if(unlikely(!ff)) return 0;
-
- ff = procfile_readall(ff);
- if(unlikely(!ff)) return 0;
-
- calls_counter++;
-
- // let ARL use this pid
- arl_ptr.p = p;
- arl_ptr.ff = ff;
-
- size_t lines = procfile_lines(ff), l;
- arl_begin(p->status_arl);
-
- for(l = 0; l < lines ;l++) {
- // debug_log("CHECK: line %zu of %zu, key '%s' = '%s'", l, lines, procfile_lineword(ff, l, 0), procfile_lineword(ff, l, 1));
- arl_ptr.line = l;
- if(unlikely(arl_check(p->status_arl,
- procfile_lineword(ff, l, 0),
- procfile_lineword(ff, l, 1)))) break;
- }
-
- p->status_vmshared = p->status_rssfile + p->status_rssshmem;
-
- // debug_log("%s uid %d, gid %d, VmSize %zu, VmRSS %zu, RssFile %zu, RssShmem %zu, shared %zu", p->comm, (int)p->uid, (int)p->gid, p->status_vmsize, p->status_vmrss, p->status_rssfile, p->status_rssshmem, p->status_vmshared);
-
- return 1;
-#endif
-}
-
-
-// ----------------------------------------------------------------------------
-
-static inline int read_proc_pid_stat(struct pid_stat *p, void *ptr) {
- (void)ptr;
-
-#ifdef __FreeBSD__
- struct kinfo_proc *proc_info = (struct kinfo_proc *)ptr;
- if (unlikely(proc_info->ki_tdflags & TDF_IDLETD))
- goto cleanup;
-#else
- static procfile *ff = NULL;
-
- if(unlikely(!p->stat_filename)) {
- char filename[FILENAME_MAX + 1];
- snprintfz(filename, FILENAME_MAX, "%s/proc/%d/stat", netdata_configured_host_prefix, p->pid);
- p->stat_filename = strdupz(filename);
- }
-
- int set_quotes = (!ff)?1:0;
-
- ff = procfile_reopen(ff, p->stat_filename, NULL, PROCFILE_FLAG_NO_ERROR_ON_FILE_IO);
- if(unlikely(!ff)) goto cleanup;
-
- // if(set_quotes) procfile_set_quotes(ff, "()");
- if(unlikely(set_quotes))
- procfile_set_open_close(ff, "(", ")");
-
- ff = procfile_readall(ff);
- if(unlikely(!ff)) goto cleanup;
-#endif
-
- p->last_stat_collected_usec = p->stat_collected_usec;
- p->stat_collected_usec = now_monotonic_usec();
- calls_counter++;
-
-#ifdef __FreeBSD__
- char *comm = proc_info->ki_comm;
- p->ppid = proc_info->ki_ppid;
-#else
- // p->pid = str2pid_t(procfile_lineword(ff, 0, 0));
- char *comm = procfile_lineword(ff, 0, 1);
- p->state = *(procfile_lineword(ff, 0, 2));
- p->ppid = (int32_t)str2pid_t(procfile_lineword(ff, 0, 3));
- // p->pgrp = (int32_t)str2pid_t(procfile_lineword(ff, 0, 4));
- // p->session = (int32_t)str2pid_t(procfile_lineword(ff, 0, 5));
- // p->tty_nr = (int32_t)str2pid_t(procfile_lineword(ff, 0, 6));
- // p->tpgid = (int32_t)str2pid_t(procfile_lineword(ff, 0, 7));
- // p->flags = str2uint64_t(procfile_lineword(ff, 0, 8));
-#endif
- if(strcmp(p->comm, comm) != 0) {
- if(unlikely(debug_enabled)) {
- if(p->comm[0])
- debug_log("\tpid %d (%s) changed name to '%s'", p->pid, p->comm, comm);
- else
- debug_log("\tJust added %d (%s)", p->pid, comm);
- }
-
- strncpyz(p->comm, comm, MAX_COMPARE_NAME);
-
- // /proc/<pid>/cmdline
- if(likely(proc_pid_cmdline_is_needed))
- managed_log(p, PID_LOG_CMDLINE, read_proc_pid_cmdline(p));
-
- assign_target_to_pid(p);
- }
-
-#ifdef __FreeBSD__
- pid_incremental_rate(stat, p->minflt, (kernel_uint_t)proc_info->ki_rusage.ru_minflt);
- pid_incremental_rate(stat, p->cminflt, (kernel_uint_t)proc_info->ki_rusage_ch.ru_minflt);
- pid_incremental_rate(stat, p->majflt, (kernel_uint_t)proc_info->ki_rusage.ru_majflt);
- pid_incremental_rate(stat, p->cmajflt, (kernel_uint_t)proc_info->ki_rusage_ch.ru_majflt);
- pid_incremental_rate(stat, p->utime, (kernel_uint_t)proc_info->ki_rusage.ru_utime.tv_sec * 100 + proc_info->ki_rusage.ru_utime.tv_usec / 10000);
- pid_incremental_rate(stat, p->stime, (kernel_uint_t)proc_info->ki_rusage.ru_stime.tv_sec * 100 + proc_info->ki_rusage.ru_stime.tv_usec / 10000);
- pid_incremental_rate(stat, p->cutime, (kernel_uint_t)proc_info->ki_rusage_ch.ru_utime.tv_sec * 100 + proc_info->ki_rusage_ch.ru_utime.tv_usec / 10000);
- pid_incremental_rate(stat, p->cstime, (kernel_uint_t)proc_info->ki_rusage_ch.ru_stime.tv_sec * 100 + proc_info->ki_rusage_ch.ru_stime.tv_usec / 10000);
-
- p->num_threads = proc_info->ki_numthreads;
-
- if(enable_guest_charts) {
- enable_guest_charts = 0;
- netdata_log_info("Guest charts aren't supported by FreeBSD");
- }
-#else
- pid_incremental_rate(stat, p->minflt, str2kernel_uint_t(procfile_lineword(ff, 0, 9)));
- pid_incremental_rate(stat, p->cminflt, str2kernel_uint_t(procfile_lineword(ff, 0, 10)));
- pid_incremental_rate(stat, p->majflt, str2kernel_uint_t(procfile_lineword(ff, 0, 11)));
- pid_incremental_rate(stat, p->cmajflt, str2kernel_uint_t(procfile_lineword(ff, 0, 12)));
- pid_incremental_rate(stat, p->utime, str2kernel_uint_t(procfile_lineword(ff, 0, 13)));
- pid_incremental_rate(stat, p->stime, str2kernel_uint_t(procfile_lineword(ff, 0, 14)));
- pid_incremental_rate(stat, p->cutime, str2kernel_uint_t(procfile_lineword(ff, 0, 15)));
- pid_incremental_rate(stat, p->cstime, str2kernel_uint_t(procfile_lineword(ff, 0, 16)));
- // p->priority = str2kernel_uint_t(procfile_lineword(ff, 0, 17));
- // p->nice = str2kernel_uint_t(procfile_lineword(ff, 0, 18));
- p->num_threads = (int32_t) str2uint32_t(procfile_lineword(ff, 0, 19), NULL);
- // p->itrealvalue = str2kernel_uint_t(procfile_lineword(ff, 0, 20));
- p->collected_starttime = str2kernel_uint_t(procfile_lineword(ff, 0, 21)) / system_hz;
- p->uptime = (global_uptime > p->collected_starttime)?(global_uptime - p->collected_starttime):0;
- // p->vsize = str2kernel_uint_t(procfile_lineword(ff, 0, 22));
- // p->rss = str2kernel_uint_t(procfile_lineword(ff, 0, 23));
- // p->rsslim = str2kernel_uint_t(procfile_lineword(ff, 0, 24));
- // p->starcode = str2kernel_uint_t(procfile_lineword(ff, 0, 25));
- // p->endcode = str2kernel_uint_t(procfile_lineword(ff, 0, 26));
- // p->startstack = str2kernel_uint_t(procfile_lineword(ff, 0, 27));
- // p->kstkesp = str2kernel_uint_t(procfile_lineword(ff, 0, 28));
- // p->kstkeip = str2kernel_uint_t(procfile_lineword(ff, 0, 29));
- // p->signal = str2kernel_uint_t(procfile_lineword(ff, 0, 30));
- // p->blocked = str2kernel_uint_t(procfile_lineword(ff, 0, 31));
- // p->sigignore = str2kernel_uint_t(procfile_lineword(ff, 0, 32));
- // p->sigcatch = str2kernel_uint_t(procfile_lineword(ff, 0, 33));
- // p->wchan = str2kernel_uint_t(procfile_lineword(ff, 0, 34));
- // p->nswap = str2kernel_uint_t(procfile_lineword(ff, 0, 35));
- // p->cnswap = str2kernel_uint_t(procfile_lineword(ff, 0, 36));
- // p->exit_signal = str2kernel_uint_t(procfile_lineword(ff, 0, 37));
- // p->processor = str2kernel_uint_t(procfile_lineword(ff, 0, 38));
- // p->rt_priority = str2kernel_uint_t(procfile_lineword(ff, 0, 39));
- // p->policy = str2kernel_uint_t(procfile_lineword(ff, 0, 40));
- // p->delayacct_blkio_ticks = str2kernel_uint_t(procfile_lineword(ff, 0, 41));
-
- if(enable_guest_charts) {
-
- pid_incremental_rate(stat, p->gtime, str2kernel_uint_t(procfile_lineword(ff, 0, 42)));
- pid_incremental_rate(stat, p->cgtime, str2kernel_uint_t(procfile_lineword(ff, 0, 43)));
-
- if (show_guest_time || p->gtime || p->cgtime) {
- p->utime -= (p->utime >= p->gtime) ? p->gtime : p->utime;
- p->cutime -= (p->cutime >= p->cgtime) ? p->cgtime : p->cutime;
- show_guest_time = 1;
- }
- }
-#endif
-
- if(unlikely(debug_enabled || (p->target && p->target->debug_enabled)))
- debug_log_int("READ PROC/PID/STAT: %s/proc/%d/stat, process: '%s' on target '%s' (dt=%llu) VALUES: utime=" KERNEL_UINT_FORMAT ", stime=" KERNEL_UINT_FORMAT ", cutime=" KERNEL_UINT_FORMAT ", cstime=" KERNEL_UINT_FORMAT ", minflt=" KERNEL_UINT_FORMAT ", majflt=" KERNEL_UINT_FORMAT ", cminflt=" KERNEL_UINT_FORMAT ", cmajflt=" KERNEL_UINT_FORMAT ", threads=%d", netdata_configured_host_prefix, p->pid, p->comm, (p->target)?p->target->name:"UNSET", p->stat_collected_usec - p->last_stat_collected_usec, p->utime, p->stime, p->cutime, p->cstime, p->minflt, p->majflt, p->cminflt, p->cmajflt, p->num_threads);
-
- if(unlikely(global_iterations_counter == 1)) {
- p->minflt = 0;
- p->cminflt = 0;
- p->majflt = 0;
- p->cmajflt = 0;
- p->utime = 0;
- p->stime = 0;
- p->gtime = 0;
- p->cutime = 0;
- p->cstime = 0;
- p->cgtime = 0;
- }
-#ifndef __FreeBSD__
- update_proc_state_count(p->state);
-#endif
- return 1;
-
-cleanup:
- p->minflt = 0;
- p->cminflt = 0;
- p->majflt = 0;
- p->cmajflt = 0;
- p->utime = 0;
- p->stime = 0;
- p->gtime = 0;
- p->cutime = 0;
- p->cstime = 0;
- p->cgtime = 0;
- p->num_threads = 0;
- // p->rss = 0;
- return 0;
-}
-
-// ----------------------------------------------------------------------------
-
-static inline int read_proc_pid_io(struct pid_stat *p, void *ptr) {
- (void)ptr;
-#ifdef __FreeBSD__
- struct kinfo_proc *proc_info = (struct kinfo_proc *)ptr;
-#else
- static procfile *ff = NULL;
-
- if(unlikely(!p->io_filename)) {
- char filename[FILENAME_MAX + 1];
- snprintfz(filename, FILENAME_MAX, "%s/proc/%d/io", netdata_configured_host_prefix, p->pid);
- p->io_filename = strdupz(filename);
- }
-
- // open the file
- ff = procfile_reopen(ff, p->io_filename, NULL, PROCFILE_FLAG_NO_ERROR_ON_FILE_IO);
- if(unlikely(!ff)) goto cleanup;
-
- ff = procfile_readall(ff);
- if(unlikely(!ff)) goto cleanup;
-#endif
-
- calls_counter++;
-
- p->last_io_collected_usec = p->io_collected_usec;
- p->io_collected_usec = now_monotonic_usec();
-
-#ifdef __FreeBSD__
- pid_incremental_rate(io, p->io_storage_bytes_read, proc_info->ki_rusage.ru_inblock);
- pid_incremental_rate(io, p->io_storage_bytes_written, proc_info->ki_rusage.ru_oublock);
-#else
- pid_incremental_rate(io, p->io_logical_bytes_read, str2kernel_uint_t(procfile_lineword(ff, 0, 1)));
- pid_incremental_rate(io, p->io_logical_bytes_written, str2kernel_uint_t(procfile_lineword(ff, 1, 1)));
- pid_incremental_rate(io, p->io_read_calls, str2kernel_uint_t(procfile_lineword(ff, 2, 1)));
- pid_incremental_rate(io, p->io_write_calls, str2kernel_uint_t(procfile_lineword(ff, 3, 1)));
- pid_incremental_rate(io, p->io_storage_bytes_read, str2kernel_uint_t(procfile_lineword(ff, 4, 1)));
- pid_incremental_rate(io, p->io_storage_bytes_written, str2kernel_uint_t(procfile_lineword(ff, 5, 1)));
- pid_incremental_rate(io, p->io_cancelled_write_bytes, str2kernel_uint_t(procfile_lineword(ff, 6, 1)));
-#endif
-
- if(unlikely(global_iterations_counter == 1)) {
- p->io_logical_bytes_read = 0;
- p->io_logical_bytes_written = 0;
- p->io_read_calls = 0;
- p->io_write_calls = 0;
- p->io_storage_bytes_read = 0;
- p->io_storage_bytes_written = 0;
- p->io_cancelled_write_bytes = 0;
- }
-
- return 1;
-
-#ifndef __FreeBSD__
-cleanup:
- p->io_logical_bytes_read = 0;
- p->io_logical_bytes_written = 0;
- p->io_read_calls = 0;
- p->io_write_calls = 0;
- p->io_storage_bytes_read = 0;
- p->io_storage_bytes_written = 0;
- p->io_cancelled_write_bytes = 0;
- return 0;
-#endif
-}
-
-#ifndef __FreeBSD__
-static inline int read_global_time() {
- static char filename[FILENAME_MAX + 1] = "";
- static procfile *ff = NULL;
- static kernel_uint_t utime_raw = 0, stime_raw = 0, gtime_raw = 0, gntime_raw = 0, ntime_raw = 0;
- static usec_t collected_usec = 0, last_collected_usec = 0;
-
- if(unlikely(!ff)) {
- snprintfz(filename, FILENAME_MAX, "%s/proc/stat", netdata_configured_host_prefix);
- ff = procfile_open(filename, " \t:", PROCFILE_FLAG_DEFAULT);
- if(unlikely(!ff)) goto cleanup;
- }
-
- ff = procfile_readall(ff);
- if(unlikely(!ff)) goto cleanup;
-
- last_collected_usec = collected_usec;
- collected_usec = now_monotonic_usec();
-
- calls_counter++;
-
- // temporary - it is added global_ntime;
- kernel_uint_t global_ntime = 0;
-
- incremental_rate(global_utime, utime_raw, str2kernel_uint_t(procfile_lineword(ff, 0, 1)), collected_usec, last_collected_usec);
- incremental_rate(global_ntime, ntime_raw, str2kernel_uint_t(procfile_lineword(ff, 0, 2)), collected_usec, last_collected_usec);
- incremental_rate(global_stime, stime_raw, str2kernel_uint_t(procfile_lineword(ff, 0, 3)), collected_usec, last_collected_usec);
- incremental_rate(global_gtime, gtime_raw, str2kernel_uint_t(procfile_lineword(ff, 0, 10)), collected_usec, last_collected_usec);
-
- global_utime += global_ntime;
-
- if(enable_guest_charts) {
- // temporary - it is added global_ntime;
- kernel_uint_t global_gntime = 0;
-
- // guest nice time, on guest time
- incremental_rate(global_gntime, gntime_raw, str2kernel_uint_t(procfile_lineword(ff, 0, 11)), collected_usec, last_collected_usec);
-
- global_gtime += global_gntime;
-
- // remove guest time from user time
- global_utime -= (global_utime > global_gtime) ? global_gtime : global_utime;
- }
-
- if(unlikely(global_iterations_counter == 1)) {
- global_utime = 0;
- global_stime = 0;
- global_gtime = 0;
- }
-
- return 1;
-
-cleanup:
- global_utime = 0;
- global_stime = 0;
- global_gtime = 0;
- return 0;
-}
-#else
-static inline int read_global_time() {
- static kernel_uint_t utime_raw = 0, stime_raw = 0, ntime_raw = 0;
- static usec_t collected_usec = 0, last_collected_usec = 0;
- long cp_time[CPUSTATES];
-
- if (unlikely(CPUSTATES != 5)) {
- goto cleanup;
- } else {
- static int mib[2] = {0, 0};
-
- if (unlikely(GETSYSCTL_SIMPLE("kern.cp_time", mib, cp_time))) {
- goto cleanup;
- }
- }
-
- last_collected_usec = collected_usec;
- collected_usec = now_monotonic_usec();
-
- calls_counter++;
-
- // temporary - it is added global_ntime;
- kernel_uint_t global_ntime = 0;
-
- incremental_rate(global_utime, utime_raw, cp_time[0] * 100LLU / system_hz, collected_usec, last_collected_usec);
- incremental_rate(global_ntime, ntime_raw, cp_time[1] * 100LLU / system_hz, collected_usec, last_collected_usec);
- incremental_rate(global_stime, stime_raw, cp_time[2] * 100LLU / system_hz, collected_usec, last_collected_usec);
-
- global_utime += global_ntime;
-
- if(unlikely(global_iterations_counter == 1)) {
- global_utime = 0;
- global_stime = 0;
- global_gtime = 0;
- }
-
- return 1;
-
-cleanup:
- global_utime = 0;
- global_stime = 0;
- global_gtime = 0;
- return 0;
-}
-#endif /* !__FreeBSD__ */
-
-// ----------------------------------------------------------------------------
-
-int file_descriptor_compare(void* a, void* b) {
-#ifdef NETDATA_INTERNAL_CHECKS
- if(((struct file_descriptor *)a)->magic != 0x0BADCAFE || ((struct file_descriptor *)b)->magic != 0x0BADCAFE)
- netdata_log_error("Corrupted index data detected. Please report this.");
-#endif /* NETDATA_INTERNAL_CHECKS */
-
- if(((struct file_descriptor *)a)->hash < ((struct file_descriptor *)b)->hash)
- return -1;
-
- else if(((struct file_descriptor *)a)->hash > ((struct file_descriptor *)b)->hash)
- return 1;
-
- else
- return strcmp(((struct file_descriptor *)a)->name, ((struct file_descriptor *)b)->name);
-}
-
-// int file_descriptor_iterator(avl_t *a) { if(a) {}; return 0; }
-
-avl_tree_type all_files_index = {
- NULL,
- file_descriptor_compare
-};
-
-static struct file_descriptor *file_descriptor_find(const char *name, uint32_t hash) {
- struct file_descriptor tmp;
- tmp.hash = (hash)?hash:simple_hash(name);
- tmp.name = name;
- tmp.count = 0;
- tmp.pos = 0;
-#ifdef NETDATA_INTERNAL_CHECKS
- tmp.magic = 0x0BADCAFE;
-#endif /* NETDATA_INTERNAL_CHECKS */
-
- return (struct file_descriptor *)avl_search(&all_files_index, (avl_t *) &tmp);
-}
-
-#define file_descriptor_add(fd) avl_insert(&all_files_index, (avl_t *)(fd))
-#define file_descriptor_remove(fd) avl_remove(&all_files_index, (avl_t *)(fd))
-
-// ----------------------------------------------------------------------------
-
-static inline void file_descriptor_not_used(int id)
-{
- if(id > 0 && id < all_files_size) {
-
-#ifdef NETDATA_INTERNAL_CHECKS
- if(all_files[id].magic != 0x0BADCAFE) {
- netdata_log_error("Ignoring request to remove empty file id %d.", id);
- return;
- }
-#endif /* NETDATA_INTERNAL_CHECKS */
-
- debug_log("decreasing slot %d (count = %d).", id, all_files[id].count);
-
- if(all_files[id].count > 0) {
- all_files[id].count--;
-
- if(!all_files[id].count) {
- debug_log(" >> slot %d is empty.", id);
-
- if(unlikely(file_descriptor_remove(&all_files[id]) != (void *)&all_files[id]))
- netdata_log_error("INTERNAL ERROR: removal of unused fd from index, removed a different fd");
-
-#ifdef NETDATA_INTERNAL_CHECKS
- all_files[id].magic = 0x00000000;
-#endif /* NETDATA_INTERNAL_CHECKS */
- all_files_len--;
- }
- }
- else
- netdata_log_error("Request to decrease counter of fd %d (%s), while the use counter is 0",
- id,
- all_files[id].name);
- }
- else
- netdata_log_error("Request to decrease counter of fd %d, which is outside the array size (1 to %d)",
- id,
- all_files_size);
-}
-
-static inline void all_files_grow() {
- void *old = all_files;
- int i;
-
- // there is no empty slot
- debug_log("extending fd array to %d entries", all_files_size + FILE_DESCRIPTORS_INCREASE_STEP);
-
- all_files = reallocz(all_files, (all_files_size + FILE_DESCRIPTORS_INCREASE_STEP) * sizeof(struct file_descriptor));
-
- // if the address changed, we have to rebuild the index
- // since all pointers are now invalid
-
- if(unlikely(old && old != (void *)all_files)) {
- debug_log(" >> re-indexing.");
-
- all_files_index.root = NULL;
- for(i = 0; i < all_files_size; i++) {
- if(!all_files[i].count) continue;
- if(unlikely(file_descriptor_add(&all_files[i]) != (void *)&all_files[i]))
- netdata_log_error("INTERNAL ERROR: duplicate indexing of fd during realloc.");
- }
-
- debug_log(" >> re-indexing done.");
- }
-
- // initialize the newly added entries
-
- for(i = all_files_size; i < (all_files_size + FILE_DESCRIPTORS_INCREASE_STEP); i++) {
- all_files[i].count = 0;
- all_files[i].name = NULL;
-#ifdef NETDATA_INTERNAL_CHECKS
- all_files[i].magic = 0x00000000;
-#endif /* NETDATA_INTERNAL_CHECKS */
- all_files[i].pos = i;
- }
-
- if(unlikely(!all_files_size)) all_files_len = 1;
- all_files_size += FILE_DESCRIPTORS_INCREASE_STEP;
-}
-
-static inline int file_descriptor_set_on_empty_slot(const char *name, uint32_t hash, FD_FILETYPE type) {
- // check we have enough memory to add it
- if(!all_files || all_files_len == all_files_size)
- all_files_grow();
-
- debug_log(" >> searching for empty slot.");
-
- // search for an empty slot
-
- static int last_pos = 0;
- int i, c;
- for(i = 0, c = last_pos ; i < all_files_size ; i++, c++) {
- if(c >= all_files_size) c = 0;
- if(c == 0) continue;
-
- if(!all_files[c].count) {
- debug_log(" >> Examining slot %d.", c);
-
-#ifdef NETDATA_INTERNAL_CHECKS
- if(all_files[c].magic == 0x0BADCAFE && all_files[c].name && file_descriptor_find(all_files[c].name, all_files[c].hash))
- netdata_log_error("fd on position %d is not cleared properly. It still has %s in it.", c, all_files[c].name);
-#endif /* NETDATA_INTERNAL_CHECKS */
-
- debug_log(" >> %s fd position %d for %s (last name: %s)", all_files[c].name?"re-using":"using", c, name, all_files[c].name);
-
- freez((void *)all_files[c].name);
- all_files[c].name = NULL;
- last_pos = c;
- break;
- }
- }
-
- all_files_len++;
-
- if(i == all_files_size) {
- fatal("We should find an empty slot, but there isn't any");
- exit(1);
- }
- // else we have an empty slot in 'c'
-
- debug_log(" >> updating slot %d.", c);
-
- all_files[c].name = strdupz(name);
- all_files[c].hash = hash;
- all_files[c].type = type;
- all_files[c].pos = c;
- all_files[c].count = 1;
-#ifdef NETDATA_INTERNAL_CHECKS
- all_files[c].magic = 0x0BADCAFE;
-#endif /* NETDATA_INTERNAL_CHECKS */
- if(unlikely(file_descriptor_add(&all_files[c]) != (void *)&all_files[c]))
- netdata_log_error("INTERNAL ERROR: duplicate indexing of fd.");
-
- debug_log("using fd position %d (name: %s)", c, all_files[c].name);
-
- return c;
-}
-
-static inline int file_descriptor_find_or_add(const char *name, uint32_t hash) {
- if(unlikely(!hash))
- hash = simple_hash(name);
-
- debug_log("adding or finding name '%s' with hash %u", name, hash);
-
- struct file_descriptor *fd = file_descriptor_find(name, hash);
- if(fd) {
- // found
- debug_log(" >> found on slot %d", fd->pos);
-
- fd->count++;
- return fd->pos;
- }
- // not found
-
- FD_FILETYPE type;
- if(likely(name[0] == '/')) type = FILETYPE_FILE;
- else if(likely(strncmp(name, "pipe:", 5) == 0)) type = FILETYPE_PIPE;
- else if(likely(strncmp(name, "socket:", 7) == 0)) type = FILETYPE_SOCKET;
- else if(likely(strncmp(name, "anon_inode:", 11) == 0)) {
- const char *t = &name[11];
-
- if(strcmp(t, "inotify") == 0) type = FILETYPE_INOTIFY;
- else if(strcmp(t, "[eventfd]") == 0) type = FILETYPE_EVENTFD;
- else if(strcmp(t, "[eventpoll]") == 0) type = FILETYPE_EVENTPOLL;
- else if(strcmp(t, "[timerfd]") == 0) type = FILETYPE_TIMERFD;
- else if(strcmp(t, "[signalfd]") == 0) type = FILETYPE_SIGNALFD;
- else {
- debug_log("UNKNOWN anonymous inode: %s", name);
- type = FILETYPE_OTHER;
- }
- }
- else if(likely(strcmp(name, "inotify") == 0)) type = FILETYPE_INOTIFY;
- else {
- debug_log("UNKNOWN linkname: %s", name);
- type = FILETYPE_OTHER;
- }
-
- return file_descriptor_set_on_empty_slot(name, hash, type);
-}
-
-static inline void clear_pid_fd(struct pid_fd *pfd) {
- pfd->fd = 0;
-
- #ifndef __FreeBSD__
- pfd->link_hash = 0;
- pfd->inode = 0;
- pfd->cache_iterations_counter = 0;
- pfd->cache_iterations_reset = 0;
-#endif
-}
-
-static inline void make_all_pid_fds_negative(struct pid_stat *p) {
- struct pid_fd *pfd = p->fds, *pfdend = &p->fds[p->fds_size];
- while(pfd < pfdend) {
- pfd->fd = -(pfd->fd);
- pfd++;
- }
-}
-
-static inline void cleanup_negative_pid_fds(struct pid_stat *p) {
- struct pid_fd *pfd = p->fds, *pfdend = &p->fds[p->fds_size];
-
- while(pfd < pfdend) {
- int fd = pfd->fd;
-
- if(unlikely(fd < 0)) {
- file_descriptor_not_used(-(fd));
- clear_pid_fd(pfd);
- }
-
- pfd++;
- }
-}
-
-static inline void init_pid_fds(struct pid_stat *p, size_t first, size_t size) {
- struct pid_fd *pfd = &p->fds[first], *pfdend = &p->fds[first + size];
-
- while(pfd < pfdend) {
-#ifndef __FreeBSD__
- pfd->filename = NULL;
-#endif
- clear_pid_fd(pfd);
- pfd++;
- }
-}
-
-static inline int read_pid_file_descriptors(struct pid_stat *p, void *ptr) {
- (void)ptr;
-#ifdef __FreeBSD__
- int mib[4];
- size_t size;
- struct kinfo_file *fds;
- static char *fdsbuf;
- char *bfdsbuf, *efdsbuf;
- char fdsname[FILENAME_MAX + 1];
-#define SHM_FORMAT_LEN 31 // format: 21 + size: 10
- char shm_name[FILENAME_MAX - SHM_FORMAT_LEN + 1];
-
- // we make all pid fds negative, so that
- // we can detect unused file descriptors
- // at the end, to free them
- make_all_pid_fds_negative(p);
-
- mib[0] = CTL_KERN;
- mib[1] = KERN_PROC;
- mib[2] = KERN_PROC_FILEDESC;
- mib[3] = p->pid;
-
- if (unlikely(sysctl(mib, 4, NULL, &size, NULL, 0))) {
- netdata_log_error("sysctl error: Can't get file descriptors data size for pid %d", p->pid);
- return 0;
- }
- if (likely(size > 0))
- fdsbuf = reallocz(fdsbuf, size);
- if (unlikely(sysctl(mib, 4, fdsbuf, &size, NULL, 0))) {
- netdata_log_error("sysctl error: Can't get file descriptors data for pid %d", p->pid);
- return 0;
- }
-
- bfdsbuf = fdsbuf;
- efdsbuf = fdsbuf + size;
- while (bfdsbuf < efdsbuf) {
- fds = (struct kinfo_file *)(uintptr_t)bfdsbuf;
- if (unlikely(fds->kf_structsize == 0))
- break;
-
- // do not process file descriptors for current working directory, root directory,
- // jail directory, ktrace vnode, text vnode and controlling terminal
- if (unlikely(fds->kf_fd < 0)) {
- bfdsbuf += fds->kf_structsize;
- continue;
- }
-
- // get file descriptors array index
- size_t fdid = fds->kf_fd;
-
- // check if the fds array is small
- if (unlikely(fdid >= p->fds_size)) {
- // it is small, extend it
-
- debug_log("extending fd memory slots for %s from %d to %d", p->comm, p->fds_size, fdid + MAX_SPARE_FDS);
-
- p->fds = reallocz(p->fds, (fdid + MAX_SPARE_FDS) * sizeof(struct pid_fd));
-
- // and initialize it
- init_pid_fds(p, p->fds_size, (fdid + MAX_SPARE_FDS) - p->fds_size);
- p->fds_size = fdid + MAX_SPARE_FDS;
- }
-
- if (unlikely(p->fds[fdid].fd == 0)) {
- // we don't know this fd, get it
-
- switch (fds->kf_type) {
- case KF_TYPE_FIFO:
- case KF_TYPE_VNODE:
- if (unlikely(!fds->kf_path[0])) {
- sprintf(fdsname, "other: inode: %lu", fds->kf_un.kf_file.kf_file_fileid);
- break;
- }
- sprintf(fdsname, "%s", fds->kf_path);
- break;
- case KF_TYPE_SOCKET:
- switch (fds->kf_sock_domain) {
- case AF_INET:
- case AF_INET6:
- if (fds->kf_sock_protocol == IPPROTO_TCP)
- sprintf(fdsname, "socket: %d %lx", fds->kf_sock_protocol, fds->kf_un.kf_sock.kf_sock_inpcb);
- else
- sprintf(fdsname, "socket: %d %lx", fds->kf_sock_protocol, fds->kf_un.kf_sock.kf_sock_pcb);
- break;
- case AF_UNIX:
- /* print address of pcb and connected pcb */
- sprintf(fdsname, "socket: %lx %lx", fds->kf_un.kf_sock.kf_sock_pcb, fds->kf_un.kf_sock.kf_sock_unpconn);
- break;
- default:
- /* print protocol number and socket address */
-#if __FreeBSD_version < 1200031
- sprintf(fdsname, "socket: other: %d %s %s", fds->kf_sock_protocol, fds->kf_sa_local.__ss_pad1, fds->kf_sa_local.__ss_pad2);
-#else
- sprintf(fdsname, "socket: other: %d %s %s", fds->kf_sock_protocol, fds->kf_un.kf_sock.kf_sa_local.__ss_pad1, fds->kf_un.kf_sock.kf_sa_local.__ss_pad2);
-#endif
- }
- break;
- case KF_TYPE_PIPE:
- sprintf(fdsname, "pipe: %lu %lu", fds->kf_un.kf_pipe.kf_pipe_addr, fds->kf_un.kf_pipe.kf_pipe_peer);
- break;
- case KF_TYPE_PTS:
-#if __FreeBSD_version < 1200031
- sprintf(fdsname, "other: pts: %u", fds->kf_un.kf_pts.kf_pts_dev);
-#else
- sprintf(fdsname, "other: pts: %lu", fds->kf_un.kf_pts.kf_pts_dev);
-#endif
- break;
- case KF_TYPE_SHM:
- strncpyz(shm_name, fds->kf_path, FILENAME_MAX - SHM_FORMAT_LEN);
- sprintf(fdsname, "other: shm: %s size: %lu", shm_name, fds->kf_un.kf_file.kf_file_size);
- break;
- case KF_TYPE_SEM:
- sprintf(fdsname, "other: sem: %u", fds->kf_un.kf_sem.kf_sem_value);
- break;
- default:
- sprintf(fdsname, "other: pid: %d fd: %d", fds->kf_un.kf_proc.kf_pid, fds->kf_fd);
- }
-
- // if another process already has this, we will get
- // the same id
- p->fds[fdid].fd = file_descriptor_find_or_add(fdsname, 0);
- }
-
- // else make it positive again, we need it
- // of course, the actual file may have changed
-
- else
- p->fds[fdid].fd = -p->fds[fdid].fd;
-
- bfdsbuf += fds->kf_structsize;
- }
-#else
- if(unlikely(!p->fds_dirname)) {
- char dirname[FILENAME_MAX+1];
- snprintfz(dirname, FILENAME_MAX, "%s/proc/%d/fd", netdata_configured_host_prefix, p->pid);
- p->fds_dirname = strdupz(dirname);
- }
-
- DIR *fds = opendir(p->fds_dirname);
- if(unlikely(!fds)) return 0;
-
- struct dirent *de;
- char linkname[FILENAME_MAX + 1];
-
- // we make all pid fds negative, so that
- // we can detect unused file descriptors
- // at the end, to free them
- make_all_pid_fds_negative(p);
-
- while((de = readdir(fds))) {
- // we need only files with numeric names
-
- if(unlikely(de->d_name[0] < '0' || de->d_name[0] > '9'))
- continue;
-
- // get its number
- int fdid = (int) str2l(de->d_name);
- if(unlikely(fdid < 0)) continue;
-
- // check if the fds array is small
- if(unlikely((size_t)fdid >= p->fds_size)) {
- // it is small, extend it
-
- debug_log("extending fd memory slots for %s from %d to %d"
- , p->comm
- , p->fds_size
- , fdid + MAX_SPARE_FDS
- );
-
- p->fds = reallocz(p->fds, (fdid + MAX_SPARE_FDS) * sizeof(struct pid_fd));
-
- // and initialize it
- init_pid_fds(p, p->fds_size, (fdid + MAX_SPARE_FDS) - p->fds_size);
- p->fds_size = (size_t)fdid + MAX_SPARE_FDS;
- }
-
- if(unlikely(p->fds[fdid].fd < 0 && de->d_ino != p->fds[fdid].inode)) {
- // inodes do not match, clear the previous entry
- inodes_changed_counter++;
- file_descriptor_not_used(-p->fds[fdid].fd);
- clear_pid_fd(&p->fds[fdid]);
- }
-
- if(p->fds[fdid].fd < 0 && p->fds[fdid].cache_iterations_counter > 0) {
- p->fds[fdid].fd = -p->fds[fdid].fd;
- p->fds[fdid].cache_iterations_counter--;
- continue;
- }
-
- if(unlikely(!p->fds[fdid].filename)) {
- filenames_allocated_counter++;
- char fdname[FILENAME_MAX + 1];
- snprintfz(fdname, FILENAME_MAX, "%s/proc/%d/fd/%s", netdata_configured_host_prefix, p->pid, de->d_name);
- p->fds[fdid].filename = strdupz(fdname);
- }
-
- file_counter++;
- ssize_t l = readlink(p->fds[fdid].filename, linkname, FILENAME_MAX);
- if(unlikely(l == -1)) {
- // cannot read the link
-
- if(debug_enabled || (p->target && p->target->debug_enabled))
- netdata_log_error("Cannot read link %s", p->fds[fdid].filename);
-
- if(unlikely(p->fds[fdid].fd < 0)) {
- file_descriptor_not_used(-p->fds[fdid].fd);
- clear_pid_fd(&p->fds[fdid]);
- }
-
- continue;
- }
- else
- linkname[l] = '\0';
-
- uint32_t link_hash = simple_hash(linkname);
-
- if(unlikely(p->fds[fdid].fd < 0 && p->fds[fdid].link_hash != link_hash)) {
- // the link changed
- links_changed_counter++;
- file_descriptor_not_used(-p->fds[fdid].fd);
- clear_pid_fd(&p->fds[fdid]);
- }
-
- if(unlikely(p->fds[fdid].fd == 0)) {
- // we don't know this fd, get it
-
- // if another process already has this, we will get
- // the same id
- p->fds[fdid].fd = file_descriptor_find_or_add(linkname, link_hash);
- p->fds[fdid].inode = de->d_ino;
- p->fds[fdid].link_hash = link_hash;
- }
- else {
- // else make it positive again, we need it
- p->fds[fdid].fd = -p->fds[fdid].fd;
- }
-
- // caching control
- // without this we read all the files on every iteration
- if(max_fds_cache_seconds > 0) {
- size_t spread = ((size_t)max_fds_cache_seconds > 10) ? 10 : (size_t)max_fds_cache_seconds;
-
- // cache it for a few iterations
- size_t max = ((size_t) max_fds_cache_seconds + (fdid % spread)) / (size_t) update_every;
- p->fds[fdid].cache_iterations_reset++;
-
- if(unlikely(p->fds[fdid].cache_iterations_reset % spread == (size_t) fdid % spread))
- p->fds[fdid].cache_iterations_reset++;
-
- if(unlikely((fdid <= 2 && p->fds[fdid].cache_iterations_reset > 5) ||
- p->fds[fdid].cache_iterations_reset > max)) {
- // for stdin, stdout, stderr (fdid <= 2) we have checked a few times, or if it goes above the max, goto max
- p->fds[fdid].cache_iterations_reset = max;
- }
-
- p->fds[fdid].cache_iterations_counter = p->fds[fdid].cache_iterations_reset;
- }
- }
-
- closedir(fds);
-#endif
- cleanup_negative_pid_fds(p);
-
- return 1;
-}
-
-// ----------------------------------------------------------------------------
-
-static inline int debug_print_process_and_parents(struct pid_stat *p, usec_t time) {
- char *prefix = "\\_ ";
- int indent = 0;
-
- if(p->parent)
- indent = debug_print_process_and_parents(p->parent, p->stat_collected_usec);
- else
- prefix = " > ";
-
- char buffer[indent + 1];
- int i;
-
- for(i = 0; i < indent ;i++) buffer[i] = ' ';
- buffer[i] = '\0';
-
- fprintf(stderr, " %s %s%s (%d %s %"PRIu64""
- , buffer
- , prefix
- , p->comm
- , p->pid
- , p->updated?"running":"exited"
- , p->stat_collected_usec - time
- );
-
- if(p->utime) fprintf(stderr, " utime=" KERNEL_UINT_FORMAT, p->utime);
- if(p->stime) fprintf(stderr, " stime=" KERNEL_UINT_FORMAT, p->stime);
- if(p->gtime) fprintf(stderr, " gtime=" KERNEL_UINT_FORMAT, p->gtime);
- if(p->cutime) fprintf(stderr, " cutime=" KERNEL_UINT_FORMAT, p->cutime);
- if(p->cstime) fprintf(stderr, " cstime=" KERNEL_UINT_FORMAT, p->cstime);
- if(p->cgtime) fprintf(stderr, " cgtime=" KERNEL_UINT_FORMAT, p->cgtime);
- if(p->minflt) fprintf(stderr, " minflt=" KERNEL_UINT_FORMAT, p->minflt);
- if(p->cminflt) fprintf(stderr, " cminflt=" KERNEL_UINT_FORMAT, p->cminflt);
- if(p->majflt) fprintf(stderr, " majflt=" KERNEL_UINT_FORMAT, p->majflt);
- if(p->cmajflt) fprintf(stderr, " cmajflt=" KERNEL_UINT_FORMAT, p->cmajflt);
- fprintf(stderr, ")\n");
-
- return indent + 1;
-}
-
-static inline void debug_print_process_tree(struct pid_stat *p, char *msg __maybe_unused) {
- debug_log("%s: process %s (%d, %s) with parents:", msg, p->comm, p->pid, p->updated?"running":"exited");
- debug_print_process_and_parents(p, p->stat_collected_usec);
-}
-
-static inline void debug_find_lost_child(struct pid_stat *pe, kernel_uint_t lost, int type) {
- int found = 0;
- struct pid_stat *p = NULL;
-
- for(p = root_of_pids; p ; p = p->next) {
- if(p == pe) continue;
-
- switch(type) {
- case 1:
- if(p->cminflt > lost) {
- fprintf(stderr, " > process %d (%s) could use the lost exited child minflt " KERNEL_UINT_FORMAT " of process %d (%s)\n", p->pid, p->comm, lost, pe->pid, pe->comm);
- found++;
- }
- break;
-
- case 2:
- if(p->cmajflt > lost) {
- fprintf(stderr, " > process %d (%s) could use the lost exited child majflt " KERNEL_UINT_FORMAT " of process %d (%s)\n", p->pid, p->comm, lost, pe->pid, pe->comm);
- found++;
- }
- break;
-
- case 3:
- if(p->cutime > lost) {
- fprintf(stderr, " > process %d (%s) could use the lost exited child utime " KERNEL_UINT_FORMAT " of process %d (%s)\n", p->pid, p->comm, lost, pe->pid, pe->comm);
- found++;
- }
- break;
-
- case 4:
- if(p->cstime > lost) {
- fprintf(stderr, " > process %d (%s) could use the lost exited child stime " KERNEL_UINT_FORMAT " of process %d (%s)\n", p->pid, p->comm, lost, pe->pid, pe->comm);
- found++;
- }
- break;
-
- case 5:
- if(p->cgtime > lost) {
- fprintf(stderr, " > process %d (%s) could use the lost exited child gtime " KERNEL_UINT_FORMAT " of process %d (%s)\n", p->pid, p->comm, lost, pe->pid, pe->comm);
- found++;
- }
- break;
- }
- }
-
- if(!found) {
- switch(type) {
- case 1:
- fprintf(stderr, " > cannot find any process to use the lost exited child minflt " KERNEL_UINT_FORMAT " of process %d (%s)\n", lost, pe->pid, pe->comm);
- break;
-
- case 2:
- fprintf(stderr, " > cannot find any process to use the lost exited child majflt " KERNEL_UINT_FORMAT " of process %d (%s)\n", lost, pe->pid, pe->comm);
- break;
-
- case 3:
- fprintf(stderr, " > cannot find any process to use the lost exited child utime " KERNEL_UINT_FORMAT " of process %d (%s)\n", lost, pe->pid, pe->comm);
- break;
-
- case 4:
- fprintf(stderr, " > cannot find any process to use the lost exited child stime " KERNEL_UINT_FORMAT " of process %d (%s)\n", lost, pe->pid, pe->comm);
- break;
-
- case 5:
- fprintf(stderr, " > cannot find any process to use the lost exited child gtime " KERNEL_UINT_FORMAT " of process %d (%s)\n", lost, pe->pid, pe->comm);
- break;
- }
- }
-}
-
-static inline kernel_uint_t remove_exited_child_from_parent(kernel_uint_t *field, kernel_uint_t *pfield) {
- kernel_uint_t absorbed = 0;
-
- if(*field > *pfield) {
- absorbed += *pfield;
- *field -= *pfield;
- *pfield = 0;
- }
- else {
- absorbed += *field;
- *pfield -= *field;
- *field = 0;
- }
-
- return absorbed;
-}
-
-static inline void process_exited_processes() {
- struct pid_stat *p;
-
- for(p = root_of_pids; p ; p = p->next) {
- if(p->updated || !p->stat_collected_usec)
- continue;
-
- kernel_uint_t utime = (p->utime_raw + p->cutime_raw) * (USEC_PER_SEC * RATES_DETAIL) / (p->stat_collected_usec - p->last_stat_collected_usec);
- kernel_uint_t stime = (p->stime_raw + p->cstime_raw) * (USEC_PER_SEC * RATES_DETAIL) / (p->stat_collected_usec - p->last_stat_collected_usec);
- kernel_uint_t gtime = (p->gtime_raw + p->cgtime_raw) * (USEC_PER_SEC * RATES_DETAIL) / (p->stat_collected_usec - p->last_stat_collected_usec);
- kernel_uint_t minflt = (p->minflt_raw + p->cminflt_raw) * (USEC_PER_SEC * RATES_DETAIL) / (p->stat_collected_usec - p->last_stat_collected_usec);
- kernel_uint_t majflt = (p->majflt_raw + p->cmajflt_raw) * (USEC_PER_SEC * RATES_DETAIL) / (p->stat_collected_usec - p->last_stat_collected_usec);
-
- if(utime + stime + gtime + minflt + majflt == 0)
- continue;
-
- if(unlikely(debug_enabled)) {
- debug_log("Absorb %s (%d %s total resources: utime=" KERNEL_UINT_FORMAT " stime=" KERNEL_UINT_FORMAT " gtime=" KERNEL_UINT_FORMAT " minflt=" KERNEL_UINT_FORMAT " majflt=" KERNEL_UINT_FORMAT ")"
- , p->comm
- , p->pid
- , p->updated?"running":"exited"
- , utime
- , stime
- , gtime
- , minflt
- , majflt
- );
- debug_print_process_tree(p, "Searching parents");
- }
-
- struct pid_stat *pp;
- for(pp = p->parent; pp ; pp = pp->parent) {
- if(!pp->updated) continue;
-
- kernel_uint_t absorbed;
- absorbed = remove_exited_child_from_parent(&utime, &pp->cutime);
- if(unlikely(debug_enabled && absorbed))
- debug_log(" > process %s (%d %s) absorbed " KERNEL_UINT_FORMAT " utime (remaining: " KERNEL_UINT_FORMAT ")", pp->comm, pp->pid, pp->updated?"running":"exited", absorbed, utime);
-
- absorbed = remove_exited_child_from_parent(&stime, &pp->cstime);
- if(unlikely(debug_enabled && absorbed))
- debug_log(" > process %s (%d %s) absorbed " KERNEL_UINT_FORMAT " stime (remaining: " KERNEL_UINT_FORMAT ")", pp->comm, pp->pid, pp->updated?"running":"exited", absorbed, stime);
-
- absorbed = remove_exited_child_from_parent(&gtime, &pp->cgtime);
- if(unlikely(debug_enabled && absorbed))
- debug_log(" > process %s (%d %s) absorbed " KERNEL_UINT_FORMAT " gtime (remaining: " KERNEL_UINT_FORMAT ")", pp->comm, pp->pid, pp->updated?"running":"exited", absorbed, gtime);
-
- absorbed = remove_exited_child_from_parent(&minflt, &pp->cminflt);
- if(unlikely(debug_enabled && absorbed))
- debug_log(" > process %s (%d %s) absorbed " KERNEL_UINT_FORMAT " minflt (remaining: " KERNEL_UINT_FORMAT ")", pp->comm, pp->pid, pp->updated?"running":"exited", absorbed, minflt);
-
- absorbed = remove_exited_child_from_parent(&majflt, &pp->cmajflt);
- if(unlikely(debug_enabled && absorbed))
- debug_log(" > process %s (%d %s) absorbed " KERNEL_UINT_FORMAT " majflt (remaining: " KERNEL_UINT_FORMAT ")", pp->comm, pp->pid, pp->updated?"running":"exited", absorbed, majflt);
- }
-
- if(unlikely(utime + stime + gtime + minflt + majflt > 0)) {
- if(unlikely(debug_enabled)) {
- if(utime) debug_find_lost_child(p, utime, 3);
- if(stime) debug_find_lost_child(p, stime, 4);
- if(gtime) debug_find_lost_child(p, gtime, 5);
- if(minflt) debug_find_lost_child(p, minflt, 1);
- if(majflt) debug_find_lost_child(p, majflt, 2);
- }
-
- p->keep = true;
-
- debug_log(" > remaining resources - KEEP - for another loop: %s (%d %s total resources: utime=" KERNEL_UINT_FORMAT " stime=" KERNEL_UINT_FORMAT " gtime=" KERNEL_UINT_FORMAT " minflt=" KERNEL_UINT_FORMAT " majflt=" KERNEL_UINT_FORMAT ")"
- , p->comm
- , p->pid
- , p->updated?"running":"exited"
- , utime
- , stime
- , gtime
- , minflt
- , majflt
- );
-
- for(pp = p->parent; pp ; pp = pp->parent) {
- if(pp->updated) break;
- pp->keep = true;
-
- debug_log(" > - KEEP - parent for another loop: %s (%d %s)"
- , pp->comm
- , pp->pid
- , pp->updated?"running":"exited"
- );
- }
-
- p->utime_raw = utime * (p->stat_collected_usec - p->last_stat_collected_usec) / (USEC_PER_SEC * RATES_DETAIL);
- p->stime_raw = stime * (p->stat_collected_usec - p->last_stat_collected_usec) / (USEC_PER_SEC * RATES_DETAIL);
- p->gtime_raw = gtime * (p->stat_collected_usec - p->last_stat_collected_usec) / (USEC_PER_SEC * RATES_DETAIL);
- p->minflt_raw = minflt * (p->stat_collected_usec - p->last_stat_collected_usec) / (USEC_PER_SEC * RATES_DETAIL);
- p->majflt_raw = majflt * (p->stat_collected_usec - p->last_stat_collected_usec) / (USEC_PER_SEC * RATES_DETAIL);
- p->cutime_raw = p->cstime_raw = p->cgtime_raw = p->cminflt_raw = p->cmajflt_raw = 0;
-
- debug_log(" ");
- }
- else
- debug_log(" > totally absorbed - DONE - %s (%d %s)"
- , p->comm
- , p->pid
- , p->updated?"running":"exited"
- );
- }
-}
-
-static inline void link_all_processes_to_their_parents(void) {
- struct pid_stat *p, *pp;
-
- // link all children to their parents
- // and update children count on parents
- for(p = root_of_pids; p ; p = p->next) {
- // for each process found
-
- p->sortlist = 0;
- p->parent = NULL;
-
- if(unlikely(!p->ppid)) {
- //unnecessary code from apps_plugin.c
- //p->parent = NULL;
- continue;
- }
-
- pp = all_pids[p->ppid];
- if(likely(pp)) {
- p->parent = pp;
- pp->children_count++;
-
- if(unlikely(debug_enabled || (p->target && p->target->debug_enabled)))
- debug_log_int("child %d (%s, %s) on target '%s' has parent %d (%s, %s). Parent: utime=" KERNEL_UINT_FORMAT ", stime=" KERNEL_UINT_FORMAT ", gtime=" KERNEL_UINT_FORMAT ", minflt=" KERNEL_UINT_FORMAT ", majflt=" KERNEL_UINT_FORMAT ", cutime=" KERNEL_UINT_FORMAT ", cstime=" KERNEL_UINT_FORMAT ", cgtime=" KERNEL_UINT_FORMAT ", cminflt=" KERNEL_UINT_FORMAT ", cmajflt=" KERNEL_UINT_FORMAT "", p->pid, p->comm, p->updated?"running":"exited", (p->target)?p->target->name:"UNSET", pp->pid, pp->comm, pp->updated?"running":"exited", pp->utime, pp->stime, pp->gtime, pp->minflt, pp->majflt, pp->cutime, pp->cstime, pp->cgtime, pp->cminflt, pp->cmajflt);
- }
- else {
- p->parent = NULL;
- netdata_log_error("pid %d %s states parent %d, but the later does not exist.", p->pid, p->comm, p->ppid);
- }
- }
-}
-
-// ----------------------------------------------------------------------------
-
-// 1. read all files in /proc
-// 2. for each numeric directory:
-// i. read /proc/pid/stat
-// ii. read /proc/pid/status
-// iii. read /proc/pid/io (requires root access)
-// iii. read the entries in directory /proc/pid/fd (requires root access)
-// for each entry:
-// a. find or create a struct file_descriptor
-// b. cleanup any old/unused file_descriptors
-
-// after all these, some pids may be linked to targets, while others may not
-
-// in case of errors, only 1 every 1000 errors is printed
-// to avoid filling up all disk space
-// if debug is enabled, all errors are printed
-
-#if (ALL_PIDS_ARE_READ_INSTANTLY == 0)
-static int compar_pid(const void *pid1, const void *pid2) {
-
- struct pid_stat *p1 = all_pids[*((pid_t *)pid1)];
- struct pid_stat *p2 = all_pids[*((pid_t *)pid2)];
-
- if(p1->sortlist > p2->sortlist)
- return -1;
- else
- return 1;
-}
-#endif
-
-static inline int collect_data_for_pid(pid_t pid, void *ptr) {
- if(unlikely(pid < 0 || pid > pid_max)) {
- netdata_log_error("Invalid pid %d read (expected %d to %d). Ignoring process.", pid, 0, pid_max);
- return 0;
- }
-
- struct pid_stat *p = get_pid_entry(pid);
- if(unlikely(!p || p->read)) return 0;
- p->read = true;
-
- // debug_log("Reading process %d (%s), sortlist %d", p->pid, p->comm, p->sortlist);
-
- // --------------------------------------------------------------------
- // /proc/<pid>/stat
-
- if(unlikely(!managed_log(p, PID_LOG_STAT, read_proc_pid_stat(p, ptr))))
- // there is no reason to proceed if we cannot get its status
- return 0;
-
- // check its parent pid
- if(unlikely(p->ppid < 0 || p->ppid > pid_max)) {
- netdata_log_error("Pid %d (command '%s') states invalid parent pid %d. Using 0.", pid, p->comm, p->ppid);
- p->ppid = 0;
- }
-
- // --------------------------------------------------------------------
- // /proc/<pid>/io
-
- managed_log(p, PID_LOG_IO, read_proc_pid_io(p, ptr));
-
- // --------------------------------------------------------------------
- // /proc/<pid>/status
-
- if(unlikely(!managed_log(p, PID_LOG_STATUS, read_proc_pid_status(p, ptr))))
- // there is no reason to proceed if we cannot get its status
- return 0;
-
- // --------------------------------------------------------------------
- // /proc/<pid>/fd
-
- if(enable_file_charts) {
- managed_log(p, PID_LOG_FDS, read_pid_file_descriptors(p, ptr));
- managed_log(p, PID_LOG_LIMITS, read_proc_pid_limits(p, ptr));
- }
-
- // --------------------------------------------------------------------
- // done!
-
- if(unlikely(debug_enabled && include_exited_childs && all_pids_count && p->ppid && all_pids[p->ppid] && !all_pids[p->ppid]->read))
- debug_log("Read process %d (%s) sortlisted %d, but its parent %d (%s) sortlisted %d, is not read", p->pid, p->comm, p->sortlist, all_pids[p->ppid]->pid, all_pids[p->ppid]->comm, all_pids[p->ppid]->sortlist);
-
- // mark it as updated
- p->updated = true;
- p->keep = false;
- p->keeploops = 0;
-
- return 1;
-}
-
-static int collect_data_for_all_processes(void) {
- struct pid_stat *p = NULL;
-
-#ifndef __FreeBSD__
- // clear process state counter
- memset(proc_state_count, 0, sizeof proc_state_count);
-#else
- int i, procnum;
-
- static size_t procbase_size = 0;
- static struct kinfo_proc *procbase = NULL;
-
- size_t new_procbase_size;
-
- int mib[3] = { CTL_KERN, KERN_PROC, KERN_PROC_PROC };
- if (unlikely(sysctl(mib, 3, NULL, &new_procbase_size, NULL, 0))) {
- netdata_log_error("sysctl error: Can't get processes data size");
- return 0;
- }
-
- // give it some air for processes that may be started
- // during this little time.
- new_procbase_size += 100 * sizeof(struct kinfo_proc);
-
- // increase the buffer if needed
- if(new_procbase_size > procbase_size) {
- procbase_size = new_procbase_size;
- procbase = reallocz(procbase, procbase_size);
- }
-
- // sysctl() gets from new_procbase_size the buffer size
- // and also returns to it the amount of data filled in
- new_procbase_size = procbase_size;
-
- // get the processes from the system
- if (unlikely(sysctl(mib, 3, procbase, &new_procbase_size, NULL, 0))) {
- netdata_log_error("sysctl error: Can't get processes data");
- return 0;
- }
-
- // based on the amount of data filled in
- // calculate the number of processes we got
- procnum = new_procbase_size / sizeof(struct kinfo_proc);
-
-#endif
-
- if(all_pids_count) {
-#if (ALL_PIDS_ARE_READ_INSTANTLY == 0)
- size_t slc = 0;
-#endif
- for(p = root_of_pids; p ; p = p->next) {
- p->read = false; // mark it as not read, so that collect_data_for_pid() will read it
- p->updated = false;
- p->merged = false;
- p->children_count = 0;
- p->parent = NULL;
-
-#if (ALL_PIDS_ARE_READ_INSTANTLY == 0)
- all_pids_sortlist[slc++] = p->pid;
-#endif
- }
-
-#if (ALL_PIDS_ARE_READ_INSTANTLY == 0)
- if(unlikely(slc != all_pids_count)) {
- netdata_log_error("Internal error: I was thinking I had %zu processes in my arrays, but it seems there are %zu.", all_pids_count, slc);
- all_pids_count = slc;
- }
-
- if(include_exited_childs) {
- // Read parents before childs
- // This is needed to prevent a situation where
- // a child is found running, but until we read
- // its parent, it has exited and its parent
- // has accumulated its resources.
-
- qsort((void *)all_pids_sortlist, (size_t)all_pids_count, sizeof(pid_t), compar_pid);
-
- // we forward read all running processes
- // collect_data_for_pid() is smart enough,
- // not to read the same pid twice per iteration
- for(slc = 0; slc < all_pids_count; slc++) {
- collect_data_for_pid(all_pids_sortlist[slc], NULL);
- }
- }
-#endif
- }
-
-#ifdef __FreeBSD__
- for (i = 0 ; i < procnum ; ++i) {
- pid_t pid = procbase[i].ki_pid;
- collect_data_for_pid(pid, &procbase[i]);
- }
-#else
- static char uptime_filename[FILENAME_MAX + 1] = "";
- if(*uptime_filename == '\0')
- snprintfz(uptime_filename, FILENAME_MAX, "%s/proc/uptime", netdata_configured_host_prefix);
-
- global_uptime = (kernel_uint_t)(uptime_msec(uptime_filename) / MSEC_PER_SEC);
-
- char dirname[FILENAME_MAX + 1];
-
- snprintfz(dirname, FILENAME_MAX, "%s/proc", netdata_configured_host_prefix);
- DIR *dir = opendir(dirname);
- if(!dir) return 0;
-
- struct dirent *de = NULL;
-
- while((de = readdir(dir))) {
- char *endptr = de->d_name;
-
- if(unlikely(de->d_type != DT_DIR || de->d_name[0] < '0' || de->d_name[0] > '9'))
- continue;
-
- pid_t pid = (pid_t) strtoul(de->d_name, &endptr, 10);
-
- // make sure we read a valid number
- if(unlikely(endptr == de->d_name || *endptr != '\0'))
- continue;
-
- collect_data_for_pid(pid, NULL);
- }
- closedir(dir);
-#endif
-
- if(!all_pids_count)
- return 0;
-
- // we need /proc/stat to normalize the cpu consumption of the exited childs
- read_global_time();
-
- // build the process tree
- link_all_processes_to_their_parents();
-
- // normally this is done
- // however we may have processes exited while we collected values
- // so let's find the exited ones
- // we do this by collecting the ownership of process
- // if we manage to get the ownership, the process still runs
- process_exited_processes();
- return 1;
-}
-
-// ----------------------------------------------------------------------------
-// update statistics on the targets
-
-// 1. link all childs to their parents
-// 2. go from bottom to top, marking as merged all childs to their parents
-// this step links all parents without a target to the child target, if any
-// 3. link all top level processes (the ones not merged) to the default target
-// 4. go from top to bottom, linking all childs without a target, to their parent target
-// after this step, all processes have a target
-// [5. for each killed pid (updated = 0), remove its usage from its target]
-// 6. zero all apps_groups_targets
-// 7. concentrate all values on the apps_groups_targets
-// 8. remove all killed processes
-// 9. find the unique file count for each target
-// check: update_apps_groups_statistics()
-
-static void cleanup_exited_pids(void) {
- size_t c;
- struct pid_stat *p = NULL;
-
- for(p = root_of_pids; p ;) {
- if(!p->updated && (!p->keep || p->keeploops > 0)) {
- if(unlikely(debug_enabled && (p->keep || p->keeploops)))
- debug_log(" > CLEANUP cannot keep exited process %d (%s) anymore - removing it.", p->pid, p->comm);
-
- for(c = 0; c < p->fds_size; c++)
- if(p->fds[c].fd > 0) {
- file_descriptor_not_used(p->fds[c].fd);
- clear_pid_fd(&p->fds[c]);
- }
-
- pid_t r = p->pid;
- p = p->next;
- del_pid_entry(r);
- }
- else {
- if(unlikely(p->keep)) p->keeploops++;
- p->keep = false;
- p = p->next;
- }
- }
-}
-
-static void apply_apps_groups_targets_inheritance(void) {
- struct pid_stat *p = NULL;
-
- // children that do not have a target
- // inherit their target from their parent
- int found = 1, loops = 0;
- while(found) {
- if(unlikely(debug_enabled)) loops++;
- found = 0;
- for(p = root_of_pids; p ; p = p->next) {
- // if this process does not have a target,
- // and it has a parent
- // and its parent has a target
- // then, set the parent's target to this process
- if(unlikely(!p->target && p->parent && p->parent->target)) {
- p->target = p->parent->target;
- found++;
-
- if(debug_enabled || (p->target && p->target->debug_enabled))
- debug_log_int("TARGET INHERITANCE: %s is inherited by %d (%s) from its parent %d (%s).", p->target->name, p->pid, p->comm, p->parent->pid, p->parent->comm);
- }
- }
- }
-
- // find all the procs with 0 childs and merge them to their parents
- // repeat, until nothing more can be done.
- int sortlist = 1;
- found = 1;
- while(found) {
- if(unlikely(debug_enabled)) loops++;
- found = 0;
-
- for(p = root_of_pids; p ; p = p->next) {
- if(unlikely(!p->sortlist && !p->children_count))
- p->sortlist = sortlist++;
-
- if(unlikely(
- !p->children_count // if this process does not have any children
- && !p->merged // and is not already merged
- && p->parent // and has a parent
- && p->parent->children_count // and its parent has children
- // and the target of this process and its parent is the same,
- // or the parent does not have a target
- && (p->target == p->parent->target || !p->parent->target)
- && p->ppid != INIT_PID // and its parent is not init
- )) {
- // mark it as merged
- p->parent->children_count--;
- p->merged = true;
-
- // the parent inherits the child's target, if it does not have a target itself
- if(unlikely(p->target && !p->parent->target)) {
- p->parent->target = p->target;
-
- if(debug_enabled || (p->target && p->target->debug_enabled))
- debug_log_int("TARGET INHERITANCE: %s is inherited by %d (%s) from its child %d (%s).", p->target->name, p->parent->pid, p->parent->comm, p->pid, p->comm);
- }
-
- found++;
- }
- }
-
- debug_log("TARGET INHERITANCE: merged %d processes", found);
- }
-
- // init goes always to default target
- if(all_pids[INIT_PID] && !all_pids[INIT_PID]->matched_by_config)
- all_pids[INIT_PID]->target = apps_groups_default_target;
-
- // pid 0 goes always to default target
- if(all_pids[0] && !all_pids[INIT_PID]->matched_by_config)
- all_pids[0]->target = apps_groups_default_target;
-
- // give a default target on all top level processes
- if(unlikely(debug_enabled)) loops++;
- for(p = root_of_pids; p ; p = p->next) {
- // if the process is not merged itself
- // then it is a top level process
- if(unlikely(!p->merged && !p->target))
- p->target = apps_groups_default_target;
-
- // make sure all processes have a sortlist
- if(unlikely(!p->sortlist))
- p->sortlist = sortlist++;
- }
-
- if(all_pids[1])
- all_pids[1]->sortlist = sortlist++;
-
- // give a target to all merged child processes
- found = 1;
- while(found) {
- if(unlikely(debug_enabled)) loops++;
- found = 0;
- for(p = root_of_pids; p ; p = p->next) {
- if(unlikely(!p->target && p->merged && p->parent && p->parent->target)) {
- p->target = p->parent->target;
- found++;
-
- if(debug_enabled || (p->target && p->target->debug_enabled))
- debug_log_int("TARGET INHERITANCE: %s is inherited by %d (%s) from its parent %d (%s) at phase 2.", p->target->name, p->pid, p->comm, p->parent->pid, p->parent->comm);
- }
- }
- }
-
- debug_log("apply_apps_groups_targets_inheritance() made %d loops on the process tree", loops);
-}
-
-static size_t zero_all_targets(struct target *root) {
- struct target *w;
- size_t count = 0;
-
- for (w = root; w ; w = w->next) {
- count++;
-
- w->minflt = 0;
- w->majflt = 0;
- w->utime = 0;
- w->stime = 0;
- w->gtime = 0;
- w->cminflt = 0;
- w->cmajflt = 0;
- w->cutime = 0;
- w->cstime = 0;
- w->cgtime = 0;
- w->num_threads = 0;
- // w->rss = 0;
- w->processes = 0;
-
- w->status_vmsize = 0;
- w->status_vmrss = 0;
- w->status_vmshared = 0;
- w->status_rssfile = 0;
- w->status_rssshmem = 0;
- w->status_vmswap = 0;
- w->status_voluntary_ctxt_switches = 0;
- w->status_nonvoluntary_ctxt_switches = 0;
-
- w->io_logical_bytes_read = 0;
- w->io_logical_bytes_written = 0;
- w->io_read_calls = 0;
- w->io_write_calls = 0;
- w->io_storage_bytes_read = 0;
- w->io_storage_bytes_written = 0;
- w->io_cancelled_write_bytes = 0;
-
- // zero file counters
- if(w->target_fds) {
- memset(w->target_fds, 0, sizeof(int) * w->target_fds_size);
- w->openfds.files = 0;
- w->openfds.pipes = 0;
- w->openfds.sockets = 0;
- w->openfds.inotifies = 0;
- w->openfds.eventfds = 0;
- w->openfds.timerfds = 0;
- w->openfds.signalfds = 0;
- w->openfds.eventpolls = 0;
- w->openfds.other = 0;
-
- w->max_open_files_percent = 0.0;
- }
-
- w->collected_starttime = 0;
- w->uptime_min = 0;
- w->uptime_sum = 0;
- w->uptime_max = 0;
-
- if(unlikely(w->root_pid)) {
- struct pid_on_target *pid_on_target_to_free, *pid_on_target = w->root_pid;
-
- while(pid_on_target) {
- pid_on_target_to_free = pid_on_target;
- pid_on_target = pid_on_target->next;
- freez(pid_on_target_to_free);
- }
-
- w->root_pid = NULL;
- }
- }
-
- return count;
-}
-
-static inline void reallocate_target_fds(struct target *w) {
- if(unlikely(!w))
- return;
-
- if(unlikely(!w->target_fds || w->target_fds_size < all_files_size)) {
- w->target_fds = reallocz(w->target_fds, sizeof(int) * all_files_size);
- memset(&w->target_fds[w->target_fds_size], 0, sizeof(int) * (all_files_size - w->target_fds_size));
- w->target_fds_size = all_files_size;
- }
-}
-
-static void aggregage_fd_type_on_openfds(FD_FILETYPE type, struct openfds *openfds) {
- switch(type) {
- case FILETYPE_FILE:
- openfds->files++;
- break;
-
- case FILETYPE_PIPE:
- openfds->pipes++;
- break;
-
- case FILETYPE_SOCKET:
- openfds->sockets++;
- break;
-
- case FILETYPE_INOTIFY:
- openfds->inotifies++;
- break;
-
- case FILETYPE_EVENTFD:
- openfds->eventfds++;
- break;
-
- case FILETYPE_TIMERFD:
- openfds->timerfds++;
- break;
-
- case FILETYPE_SIGNALFD:
- openfds->signalfds++;
- break;
-
- case FILETYPE_EVENTPOLL:
- openfds->eventpolls++;
- break;
-
- case FILETYPE_OTHER:
- openfds->other++;
- break;
- }
-}
-
-static inline void aggregate_fd_on_target(int fd, struct target *w) {
- if(unlikely(!w))
- return;
-
- if(unlikely(w->target_fds[fd])) {
- // it is already aggregated
- // just increase its usage counter
- w->target_fds[fd]++;
- return;
- }
-
- // increase its usage counter
- // so that we will not add it again
- w->target_fds[fd]++;
-
- aggregage_fd_type_on_openfds(all_files[fd].type, &w->openfds);
-}
-
-static inline void aggregate_pid_fds_on_targets(struct pid_stat *p) {
-
- if(unlikely(!p->updated)) {
- // the process is not running
- return;
- }
-
- struct target *w = p->target, *u = p->user_target, *g = p->group_target;
-
- reallocate_target_fds(w);
- reallocate_target_fds(u);
- reallocate_target_fds(g);
-
- p->openfds.files = 0;
- p->openfds.pipes = 0;
- p->openfds.sockets = 0;
- p->openfds.inotifies = 0;
- p->openfds.eventfds = 0;
- p->openfds.timerfds = 0;
- p->openfds.signalfds = 0;
- p->openfds.eventpolls = 0;
- p->openfds.other = 0;
-
- long currentfds = 0;
- size_t c, size = p->fds_size;
- struct pid_fd *fds = p->fds;
- for(c = 0; c < size ;c++) {
- int fd = fds[c].fd;
-
- if(likely(fd <= 0 || fd >= all_files_size))
- continue;
-
- currentfds++;
- aggregage_fd_type_on_openfds(all_files[fd].type, &p->openfds);
-
- aggregate_fd_on_target(fd, w);
- aggregate_fd_on_target(fd, u);
- aggregate_fd_on_target(fd, g);
- }
-}
-
-static inline void aggregate_pid_on_target(struct target *w, struct pid_stat *p, struct target *o) {
- (void)o;
-
- if(unlikely(!p->updated)) {
- // the process is not running
- return;
- }
-
- if(unlikely(!w)) {
- netdata_log_error("pid %d %s was left without a target!", p->pid, p->comm);
- return;
- }
-
- if(p->openfds_limits_percent > w->max_open_files_percent)
- w->max_open_files_percent = p->openfds_limits_percent;
-
- w->cutime += p->cutime;
- w->cstime += p->cstime;
- w->cgtime += p->cgtime;
- w->cminflt += p->cminflt;
- w->cmajflt += p->cmajflt;
-
- w->utime += p->utime;
- w->stime += p->stime;
- w->gtime += p->gtime;
- w->minflt += p->minflt;
- w->majflt += p->majflt;
-
- // w->rss += p->rss;
-
- w->status_vmsize += p->status_vmsize;
- w->status_vmrss += p->status_vmrss;
- w->status_vmshared += p->status_vmshared;
- w->status_rssfile += p->status_rssfile;
- w->status_rssshmem += p->status_rssshmem;
- w->status_vmswap += p->status_vmswap;
- w->status_voluntary_ctxt_switches += p->status_voluntary_ctxt_switches;
- w->status_nonvoluntary_ctxt_switches += p->status_nonvoluntary_ctxt_switches;
-
- w->io_logical_bytes_read += p->io_logical_bytes_read;
- w->io_logical_bytes_written += p->io_logical_bytes_written;
- w->io_read_calls += p->io_read_calls;
- w->io_write_calls += p->io_write_calls;
- w->io_storage_bytes_read += p->io_storage_bytes_read;
- w->io_storage_bytes_written += p->io_storage_bytes_written;
- w->io_cancelled_write_bytes += p->io_cancelled_write_bytes;
-
- w->processes++;
- w->num_threads += p->num_threads;
-
- if(!w->collected_starttime || p->collected_starttime < w->collected_starttime) w->collected_starttime = p->collected_starttime;
- if(!w->uptime_min || p->uptime < w->uptime_min) w->uptime_min = p->uptime;
- w->uptime_sum += p->uptime;
- if(!w->uptime_max || w->uptime_max < p->uptime) w->uptime_max = p->uptime;
-
- if(unlikely(debug_enabled || w->debug_enabled)) {
- debug_log_int("aggregating '%s' pid %d on target '%s' utime=" KERNEL_UINT_FORMAT ", stime=" KERNEL_UINT_FORMAT ", gtime=" KERNEL_UINT_FORMAT ", cutime=" KERNEL_UINT_FORMAT ", cstime=" KERNEL_UINT_FORMAT ", cgtime=" KERNEL_UINT_FORMAT ", minflt=" KERNEL_UINT_FORMAT ", majflt=" KERNEL_UINT_FORMAT ", cminflt=" KERNEL_UINT_FORMAT ", cmajflt=" KERNEL_UINT_FORMAT "", p->comm, p->pid, w->name, p->utime, p->stime, p->gtime, p->cutime, p->cstime, p->cgtime, p->minflt, p->majflt, p->cminflt, p->cmajflt);
-
- struct pid_on_target *pid_on_target = mallocz(sizeof(struct pid_on_target));
- pid_on_target->pid = p->pid;
- pid_on_target->next = w->root_pid;
- w->root_pid = pid_on_target;
- }
-}
-
-static inline void post_aggregate_targets(struct target *root) {
- struct target *w;
- for (w = root; w ; w = w->next) {
- if(w->collected_starttime) {
- if (!w->starttime || w->collected_starttime < w->starttime) {
- w->starttime = w->collected_starttime;
- }
- } else {
- w->starttime = 0;
- }
- }
-}
-
-static void calculate_netdata_statistics(void) {
-
- apply_apps_groups_targets_inheritance();
-
- zero_all_targets(users_root_target);
- zero_all_targets(groups_root_target);
- apps_groups_targets_count = zero_all_targets(apps_groups_root_target);
-
- // this has to be done, before the cleanup
- struct pid_stat *p = NULL;
- struct target *w = NULL, *o = NULL;
-
- // concentrate everything on the targets
- for(p = root_of_pids; p ; p = p->next) {
-
- // --------------------------------------------------------------------
- // apps_groups target
-
- aggregate_pid_on_target(p->target, p, NULL);
-
-
- // --------------------------------------------------------------------
- // user target
-
- o = p->user_target;
- if(likely(p->user_target && p->user_target->uid == p->uid))
- w = p->user_target;
- else {
- if(unlikely(debug_enabled && p->user_target))
- debug_log("pid %d (%s) switched user from %u (%s) to %u.", p->pid, p->comm, p->user_target->uid, p->user_target->name, p->uid);
-
- w = p->user_target = get_users_target(p->uid);
- }
-
- aggregate_pid_on_target(w, p, o);
-
-
- // --------------------------------------------------------------------
- // user group target
-
- o = p->group_target;
- if(likely(p->group_target && p->group_target->gid == p->gid))
- w = p->group_target;
- else {
- if(unlikely(debug_enabled && p->group_target))
- debug_log("pid %d (%s) switched group from %u (%s) to %u.", p->pid, p->comm, p->group_target->gid, p->group_target->name, p->gid);
-
- w = p->group_target = get_groups_target(p->gid);
- }
-
- aggregate_pid_on_target(w, p, o);
-
-
- // --------------------------------------------------------------------
- // aggregate all file descriptors
-
- if(enable_file_charts)
- aggregate_pid_fds_on_targets(p);
- }
-
- post_aggregate_targets(apps_groups_root_target);
- post_aggregate_targets(users_root_target);
- post_aggregate_targets(groups_root_target);
-
- cleanup_exited_pids();
-}
-
-// ----------------------------------------------------------------------------
-// update chart dimensions
-
-static inline void send_BEGIN(const char *type, const char *name,const char *metric, usec_t usec) {
- fprintf(stdout, "BEGIN %s.%s_%s %" PRIu64 "\n", type, name, metric, usec);
-}
-
-static inline void send_SET(const char *name, kernel_uint_t value) {
- fprintf(stdout, "SET %s = " KERNEL_UINT_FORMAT "\n", name, value);
-}
-
-static inline void send_END(void) {
- fprintf(stdout, "END\n\n");
-}
-
-void send_resource_usage_to_netdata(usec_t dt) {
- static struct timeval last = { 0, 0 };
- static struct rusage me_last;
-
- struct timeval now;
- struct rusage me;
-
- usec_t cpuuser;
- usec_t cpusyst;
-
- if(!last.tv_sec) {
- now_monotonic_timeval(&last);
- getrusage(RUSAGE_SELF, &me_last);
-
- cpuuser = 0;
- cpusyst = 0;
- }
- else {
- now_monotonic_timeval(&now);
- getrusage(RUSAGE_SELF, &me);
-
- cpuuser = me.ru_utime.tv_sec * USEC_PER_SEC + me.ru_utime.tv_usec;
- cpusyst = me.ru_stime.tv_sec * USEC_PER_SEC + me.ru_stime.tv_usec;
-
- memmove(&last, &now, sizeof(struct timeval));
- memmove(&me_last, &me, sizeof(struct rusage));
- }
-
- static char created_charts = 0;
- if(unlikely(!created_charts)) {
- created_charts = 1;
-
- fprintf(stdout,
- "CHART netdata.apps_cpu '' 'Apps Plugin CPU' 'milliseconds/s' apps.plugin netdata.apps_cpu stacked 140000 %1$d\n"
- "DIMENSION user '' incremental 1 1000\n"
- "DIMENSION system '' incremental 1 1000\n"
- "CHART netdata.apps_sizes '' 'Apps Plugin Files' 'files/s' apps.plugin netdata.apps_sizes line 140001 %1$d\n"
- "DIMENSION calls '' incremental 1 1\n"
- "DIMENSION files '' incremental 1 1\n"
- "DIMENSION filenames '' incremental 1 1\n"
- "DIMENSION inode_changes '' incremental 1 1\n"
- "DIMENSION link_changes '' incremental 1 1\n"
- "DIMENSION pids '' absolute 1 1\n"
- "DIMENSION fds '' absolute 1 1\n"
- "DIMENSION targets '' absolute 1 1\n"
- "DIMENSION new_pids 'new pids' incremental 1 1\n"
- , update_every
- );
-
- fprintf(stdout,
- "CHART netdata.apps_fix '' 'Apps Plugin Normalization Ratios' 'percentage' apps.plugin netdata.apps_fix line 140002 %1$d\n"
- "DIMENSION utime '' absolute 1 %2$llu\n"
- "DIMENSION stime '' absolute 1 %2$llu\n"
- "DIMENSION gtime '' absolute 1 %2$llu\n"
- "DIMENSION minflt '' absolute 1 %2$llu\n"
- "DIMENSION majflt '' absolute 1 %2$llu\n"
- , update_every
- , RATES_DETAIL
- );
-
- if(include_exited_childs)
- fprintf(stdout,
- "CHART netdata.apps_children_fix '' 'Apps Plugin Exited Children Normalization Ratios' 'percentage' apps.plugin netdata.apps_children_fix line 140003 %1$d\n"
- "DIMENSION cutime '' absolute 1 %2$llu\n"
- "DIMENSION cstime '' absolute 1 %2$llu\n"
- "DIMENSION cgtime '' absolute 1 %2$llu\n"
- "DIMENSION cminflt '' absolute 1 %2$llu\n"
- "DIMENSION cmajflt '' absolute 1 %2$llu\n"
- , update_every
- , RATES_DETAIL
- );
-
- }
-
- fprintf(stdout,
- "BEGIN netdata.apps_cpu %"PRIu64"\n"
- "SET user = %"PRIu64"\n"
- "SET system = %"PRIu64"\n"
- "END\n"
- "BEGIN netdata.apps_sizes %"PRIu64"\n"
- "SET calls = %zu\n"
- "SET files = %zu\n"
- "SET filenames = %zu\n"
- "SET inode_changes = %zu\n"
- "SET link_changes = %zu\n"
- "SET pids = %zu\n"
- "SET fds = %d\n"
- "SET targets = %zu\n"
- "SET new_pids = %zu\n"
- "END\n"
- , dt
- , cpuuser
- , cpusyst
- , dt
- , calls_counter
- , file_counter
- , filenames_allocated_counter
- , inodes_changed_counter
- , links_changed_counter
- , all_pids_count
- , all_files_len
- , apps_groups_targets_count
- , targets_assignment_counter
- );
-
- fprintf(stdout,
- "BEGIN netdata.apps_fix %"PRIu64"\n"
- "SET utime = %u\n"
- "SET stime = %u\n"
- "SET gtime = %u\n"
- "SET minflt = %u\n"
- "SET majflt = %u\n"
- "END\n"
- , dt
- , (unsigned int)(utime_fix_ratio * 100 * RATES_DETAIL)
- , (unsigned int)(stime_fix_ratio * 100 * RATES_DETAIL)
- , (unsigned int)(gtime_fix_ratio * 100 * RATES_DETAIL)
- , (unsigned int)(minflt_fix_ratio * 100 * RATES_DETAIL)
- , (unsigned int)(majflt_fix_ratio * 100 * RATES_DETAIL)
- );
-
- if(include_exited_childs)
- fprintf(stdout,
- "BEGIN netdata.apps_children_fix %"PRIu64"\n"
- "SET cutime = %u\n"
- "SET cstime = %u\n"
- "SET cgtime = %u\n"
- "SET cminflt = %u\n"
- "SET cmajflt = %u\n"
- "END\n"
- , dt
- , (unsigned int)(cutime_fix_ratio * 100 * RATES_DETAIL)
- , (unsigned int)(cstime_fix_ratio * 100 * RATES_DETAIL)
- , (unsigned int)(cgtime_fix_ratio * 100 * RATES_DETAIL)
- , (unsigned int)(cminflt_fix_ratio * 100 * RATES_DETAIL)
- , (unsigned int)(cmajflt_fix_ratio * 100 * RATES_DETAIL)
- );
-}
-
-static void normalize_utilization(struct target *root) {
- struct target *w;
-
- // childs processing introduces spikes
- // here we try to eliminate them by disabling childs processing either for specific dimensions
- // or entirely. Of course, either way, we disable it just a single iteration.
-
- kernel_uint_t max_time = get_system_cpus() * time_factor * RATES_DETAIL;
- kernel_uint_t utime = 0, cutime = 0, stime = 0, cstime = 0, gtime = 0, cgtime = 0, minflt = 0, cminflt = 0, majflt = 0, cmajflt = 0;
-
- if(global_utime > max_time) global_utime = max_time;
- if(global_stime > max_time) global_stime = max_time;
- if(global_gtime > max_time) global_gtime = max_time;
-
- for(w = root; w ; w = w->next) {
- if(w->target || (!w->processes && !w->exposed)) continue;
-
- utime += w->utime;
- stime += w->stime;
- gtime += w->gtime;
- cutime += w->cutime;
- cstime += w->cstime;
- cgtime += w->cgtime;
-
- minflt += w->minflt;
- majflt += w->majflt;
- cminflt += w->cminflt;
- cmajflt += w->cmajflt;
- }
-
- if(global_utime || global_stime || global_gtime) {
- if(global_utime + global_stime + global_gtime > utime + cutime + stime + cstime + gtime + cgtime) {
- // everything we collected fits
- utime_fix_ratio =
- stime_fix_ratio =
- gtime_fix_ratio =
- cutime_fix_ratio =
- cstime_fix_ratio =
- cgtime_fix_ratio = 1.0; //(NETDATA_DOUBLE)(global_utime + global_stime) / (NETDATA_DOUBLE)(utime + cutime + stime + cstime);
- }
- else if((global_utime + global_stime > utime + stime) && (cutime || cstime)) {
- // children resources are too high
- // lower only the children resources
- utime_fix_ratio =
- stime_fix_ratio =
- gtime_fix_ratio = 1.0;
- cutime_fix_ratio =
- cstime_fix_ratio =
- cgtime_fix_ratio = (NETDATA_DOUBLE)((global_utime + global_stime) - (utime + stime)) / (NETDATA_DOUBLE)(cutime + cstime);
- }
- else if(utime || stime) {
- // even running processes are unrealistic
- // zero the children resources
- // lower the running processes resources
- utime_fix_ratio =
- stime_fix_ratio =
- gtime_fix_ratio = (NETDATA_DOUBLE)(global_utime + global_stime) / (NETDATA_DOUBLE)(utime + stime);
- cutime_fix_ratio =
- cstime_fix_ratio =
- cgtime_fix_ratio = 0.0;
- }
- else {
- utime_fix_ratio =
- stime_fix_ratio =
- gtime_fix_ratio =
- cutime_fix_ratio =
- cstime_fix_ratio =
- cgtime_fix_ratio = 0.0;
- }
- }
- else {
- utime_fix_ratio =
- stime_fix_ratio =
- gtime_fix_ratio =
- cutime_fix_ratio =
- cstime_fix_ratio =
- cgtime_fix_ratio = 0.0;
- }
-
- if(utime_fix_ratio > 1.0) utime_fix_ratio = 1.0;
- if(cutime_fix_ratio > 1.0) cutime_fix_ratio = 1.0;
- if(stime_fix_ratio > 1.0) stime_fix_ratio = 1.0;
- if(cstime_fix_ratio > 1.0) cstime_fix_ratio = 1.0;
- if(gtime_fix_ratio > 1.0) gtime_fix_ratio = 1.0;
- if(cgtime_fix_ratio > 1.0) cgtime_fix_ratio = 1.0;
-
- // if(utime_fix_ratio < 0.0) utime_fix_ratio = 0.0;
- // if(cutime_fix_ratio < 0.0) cutime_fix_ratio = 0.0;
- // if(stime_fix_ratio < 0.0) stime_fix_ratio = 0.0;
- // if(cstime_fix_ratio < 0.0) cstime_fix_ratio = 0.0;
- // if(gtime_fix_ratio < 0.0) gtime_fix_ratio = 0.0;
- // if(cgtime_fix_ratio < 0.0) cgtime_fix_ratio = 0.0;
-
- // TODO
- // we use cpu time to normalize page faults
- // the problem is that to find the proper max values
- // for page faults we have to parse /proc/vmstat
- // which is quite big to do it again (netdata does it already)
- //
- // a better solution could be to somehow have netdata
- // do this normalization for us
-
- if(utime || stime || gtime)
- majflt_fix_ratio =
- minflt_fix_ratio = (NETDATA_DOUBLE)(utime * utime_fix_ratio + stime * stime_fix_ratio + gtime * gtime_fix_ratio) / (NETDATA_DOUBLE)(utime + stime + gtime);
- else
- minflt_fix_ratio =
- majflt_fix_ratio = 1.0;
-
- if(cutime || cstime || cgtime)
- cmajflt_fix_ratio =
- cminflt_fix_ratio = (NETDATA_DOUBLE)(cutime * cutime_fix_ratio + cstime * cstime_fix_ratio + cgtime * cgtime_fix_ratio) / (NETDATA_DOUBLE)(cutime + cstime + cgtime);
- else
- cminflt_fix_ratio =
- cmajflt_fix_ratio = 1.0;
-
- // the report
-
- debug_log(
- "SYSTEM: u=" KERNEL_UINT_FORMAT " s=" KERNEL_UINT_FORMAT " g=" KERNEL_UINT_FORMAT " "
- "COLLECTED: u=" KERNEL_UINT_FORMAT " s=" KERNEL_UINT_FORMAT " g=" KERNEL_UINT_FORMAT " cu=" KERNEL_UINT_FORMAT " cs=" KERNEL_UINT_FORMAT " cg=" KERNEL_UINT_FORMAT " "
- "DELTA: u=" KERNEL_UINT_FORMAT " s=" KERNEL_UINT_FORMAT " g=" KERNEL_UINT_FORMAT " "
- "FIX: u=%0.2f s=%0.2f g=%0.2f cu=%0.2f cs=%0.2f cg=%0.2f "
- "FINALLY: u=" KERNEL_UINT_FORMAT " s=" KERNEL_UINT_FORMAT " g=" KERNEL_UINT_FORMAT " cu=" KERNEL_UINT_FORMAT " cs=" KERNEL_UINT_FORMAT " cg=" KERNEL_UINT_FORMAT " "
- , global_utime
- , global_stime
- , global_gtime
- , utime
- , stime
- , gtime
- , cutime
- , cstime
- , cgtime
- , utime + cutime - global_utime
- , stime + cstime - global_stime
- , gtime + cgtime - global_gtime
- , utime_fix_ratio
- , stime_fix_ratio
- , gtime_fix_ratio
- , cutime_fix_ratio
- , cstime_fix_ratio
- , cgtime_fix_ratio
- , (kernel_uint_t)(utime * utime_fix_ratio)
- , (kernel_uint_t)(stime * stime_fix_ratio)
- , (kernel_uint_t)(gtime * gtime_fix_ratio)
- , (kernel_uint_t)(cutime * cutime_fix_ratio)
- , (kernel_uint_t)(cstime * cstime_fix_ratio)
- , (kernel_uint_t)(cgtime * cgtime_fix_ratio)
- );
-}
-
-static void send_collected_data_to_netdata(struct target *root, const char *type, usec_t dt) {
- struct target *w;
-
- for (w = root; w ; w = w->next) {
- if (unlikely(!w->exposed))
- continue;
-
- send_BEGIN(type, w->clean_name, "processes", dt);
- send_SET("processes", w->processes);
- send_END();
-
- send_BEGIN(type, w->clean_name, "threads", dt);
- send_SET("threads", w->num_threads);
- send_END();
-
- if (unlikely(!w->processes && !w->is_other))
- continue;
-
- send_BEGIN(type, w->clean_name, "cpu_utilization", dt);
- send_SET("user", (kernel_uint_t)(w->utime * utime_fix_ratio) + (include_exited_childs ? ((kernel_uint_t)(w->cutime * cutime_fix_ratio)) : 0ULL));
- send_SET("system", (kernel_uint_t)(w->stime * stime_fix_ratio) + (include_exited_childs ? ((kernel_uint_t)(w->cstime * cstime_fix_ratio)) : 0ULL));
- send_END();
-
-#ifndef __FreeBSD__
- if (enable_guest_charts) {
- send_BEGIN(type, w->clean_name, "cpu_guest_utilization", dt);
- send_SET("guest", (kernel_uint_t)(w->gtime * gtime_fix_ratio) + (include_exited_childs ? ((kernel_uint_t)(w->cgtime * cgtime_fix_ratio)) : 0ULL));
- send_END();
- }
-
- send_BEGIN(type, w->clean_name, "cpu_context_switches", dt);
- send_SET("voluntary", w->status_voluntary_ctxt_switches);
- send_SET("involuntary", w->status_nonvoluntary_ctxt_switches);
- send_END();
-
- send_BEGIN(type, w->clean_name, "mem_private_usage", dt);
- send_SET("mem", (w->status_vmrss > w->status_vmshared)?(w->status_vmrss - w->status_vmshared) : 0ULL);
- send_END();
-#endif
-
- send_BEGIN(type, w->clean_name, "mem_usage", dt);
- send_SET("rss", w->status_vmrss);
- send_END();
-
- send_BEGIN(type, w->clean_name, "vmem_usage", dt);
- send_SET("vmem", w->status_vmsize);
- send_END();
-
- send_BEGIN(type, w->clean_name, "mem_page_faults", dt);
- send_SET("minor", (kernel_uint_t)(w->minflt * minflt_fix_ratio) + (include_exited_childs ? ((kernel_uint_t)(w->cminflt * cminflt_fix_ratio)) : 0ULL));
- send_SET("major", (kernel_uint_t)(w->majflt * majflt_fix_ratio) + (include_exited_childs ? ((kernel_uint_t)(w->cmajflt * cmajflt_fix_ratio)) : 0ULL));
- send_END();
-
-#ifndef __FreeBSD__
- send_BEGIN(type, w->clean_name, "swap_usage", dt);
- send_SET("swap", w->status_vmswap);
- send_END();
-#endif
-
-#ifndef __FreeBSD__
- if (w->processes == 0) {
- send_BEGIN(type, w->clean_name, "uptime", dt);
- send_SET("uptime", 0);
- send_END();
-
- if (enable_detailed_uptime_charts) {
- send_BEGIN(type, w->clean_name, "uptime_summary", dt);
- send_SET("min", 0);
- send_SET("avg", 0);
- send_SET("max", 0);
- send_END();
- }
- } else {
- send_BEGIN(type, w->clean_name, "uptime", dt);
- send_SET("uptime", (global_uptime > w->starttime) ? (global_uptime - w->starttime) : 0);
- send_END();
-
- if (enable_detailed_uptime_charts) {
- send_BEGIN(type, w->clean_name, "uptime_summary", dt);
- send_SET("min", w->uptime_min);
- send_SET("avg", w->processes > 0 ? w->uptime_sum / w->processes : 0);
- send_SET("max", w->uptime_max);
- send_END();
- }
- }
-#endif
-
- send_BEGIN(type, w->clean_name, "disk_physical_io", dt);
- send_SET("reads", w->io_storage_bytes_read);
- send_SET("writes", w->io_storage_bytes_written);
- send_END();
-
-#ifndef __FreeBSD__
- send_BEGIN(type, w->clean_name, "disk_logical_io", dt);
- send_SET("reads", w->io_logical_bytes_read);
- send_SET("writes", w->io_logical_bytes_written);
- send_END();
-#endif
- if (enable_file_charts) {
- send_BEGIN(type, w->clean_name, "fds_open_limit", dt);
- send_SET("limit", w->max_open_files_percent * 100.0);
- send_END();
-
- send_BEGIN(type, w->clean_name, "fds_open", dt);
- send_SET("files", w->openfds.files);
- send_SET("sockets", w->openfds.sockets);
- send_SET("pipes", w->openfds.sockets);
- send_SET("inotifies", w->openfds.inotifies);
- send_SET("event", w->openfds.eventfds);
- send_SET("timer", w->openfds.timerfds);
- send_SET("signal", w->openfds.signalfds);
- send_SET("eventpolls", w->openfds.eventpolls);
- send_SET("other", w->openfds.other);
- send_END();
- }
- }
-}
-
-
-// ----------------------------------------------------------------------------
-// generate the charts
-
-static void send_charts_updates_to_netdata(struct target *root, const char *type, const char *lbl_name, const char *title)
-{
- struct target *w;
-
- if (debug_enabled) {
- for (w = root; w; w = w->next) {
- if (unlikely(!w->target && w->processes)) {
- struct pid_on_target *pid_on_target;
- fprintf(stderr, "apps.plugin: target '%s' has aggregated %u process(es):", w->name, w->processes);
- for (pid_on_target = w->root_pid; pid_on_target; pid_on_target = pid_on_target->next) {
- fprintf(stderr, " %d", pid_on_target->pid);
- }
- fputc('\n', stderr);
- }
- }
- }
-
- for (w = root; w; w = w->next) {
- if (likely(w->exposed || (!w->processes && !w->is_other)))
- continue;
-
- w->exposed = 1;
-
- fprintf(stdout, "CHART %s.%s_cpu_utilization '' '%s CPU utilization (100%% = 1 core)' 'percentage' cpu %s.cpu_utilization stacked 20001 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION user '' absolute 1 %llu\n", time_factor * RATES_DETAIL / 100LLU);
- fprintf(stdout, "DIMENSION system '' absolute 1 %llu\n", time_factor * RATES_DETAIL / 100LLU);
-
-#ifndef __FreeBSD__
- if (enable_guest_charts) {
- fprintf(stdout, "CHART %s.%s_cpu_guest_utilization '' '%s CPU guest utlization (100%% = 1 core)' 'percentage' cpu %s.cpu_guest_utilization line 20005 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION guest '' absolute 1 %llu\n", time_factor * RATES_DETAIL / 100LLU);
- }
-
- fprintf(stdout, "CHART %s.%s_cpu_context_switches '' '%s CPU context switches' 'switches/s' cpu %s.cpu_context_switches stacked 20010 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION voluntary '' absolute 1 %llu\n", RATES_DETAIL);
- fprintf(stdout, "DIMENSION involuntary '' absolute 1 %llu\n", RATES_DETAIL);
-
- fprintf(stdout, "CHART %s.%s_mem_private_usage '' '%s memory usage without shared' 'MiB' mem %s.mem_private_usage area 20050 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION mem '' absolute %ld %ld\n", 1L, 1024L);
-#endif
-
- fprintf(stdout, "CHART %s.%s_mem_usage '' '%s memory RSS usage' 'MiB' mem %s.mem_usage area 20055 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION rss '' absolute %ld %ld\n", 1L, 1024L);
-
- fprintf(stdout, "CHART %s.%s_mem_page_faults '' '%s memory page faults' 'pgfaults/s' mem %s.mem_page_faults stacked 20060 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION major '' absolute 1 %llu\n", RATES_DETAIL);
- fprintf(stdout, "DIMENSION minor '' absolute 1 %llu\n", RATES_DETAIL);
-
- fprintf(stdout, "CHART %s.%s_vmem_usage '' '%s virtual memory size' 'MiB' mem %s.vmem_usage line 20065 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION vmem '' absolute %ld %ld\n", 1L, 1024L);
-
-#ifndef __FreeBSD__
- fprintf(stdout, "CHART %s.%s_swap_usage '' '%s swap usage' 'MiB' mem %s.swap_usage area 20065 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION swap '' absolute %ld %ld\n", 1L, 1024L);
-#endif
-
-#ifndef __FreeBSD__
- fprintf(stdout, "CHART %s.%s_disk_physical_io '' '%s disk physical IO' 'KiB/s' disk %s.disk_physical_io area 20100 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION reads '' absolute 1 %llu\n", 1024LLU * RATES_DETAIL);
- fprintf(stdout, "DIMENSION writes '' absolute -1 %llu\n", 1024LLU * RATES_DETAIL);
-
- fprintf(stdout, "CHART %s.%s_disk_logical_io '' '%s disk logical IO' 'KiB/s' disk %s.disk_logical_io area 20105 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION reads '' absolute 1 %llu\n", 1024LLU * RATES_DETAIL);
- fprintf(stdout, "DIMENSION writes '' absolute -1 %llu\n", 1024LLU * RATES_DETAIL);
-#else
- fprintf(stdout, "CHART %s.%s_disk_physical_io '' '%s disk physical IO' 'blocks/s' disk %s.disk_physical_block_io area 20100 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION reads '' absolute 1 %llu\n", RATES_DETAIL);
- fprintf(stdout, "DIMENSION writes '' absolute -1 %llu\n", RATES_DETAIL);
-#endif
-
- fprintf(stdout, "CHART %s.%s_processes '' '%s processes' 'processes' processes %s.processes line 20150 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION processes '' absolute 1 1\n");
-
- fprintf(stdout, "CHART %s.%s_threads '' '%s threads' 'threads' processes %s.threads line 20155 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION threads '' absolute 1 1\n");
-
- if (enable_file_charts) {
- fprintf(stdout, "CHART %s.%s_fds_open_limit '' '%s open file descriptors limit' '%%' fds %s.fds_open_limit line 20200 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION limit '' absolute 1 100\n");
-
- fprintf(stdout, "CHART %s.%s_fds_open '' '%s open files descriptors' 'fds' fds %s.fds_open stacked 20210 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION files '' absolute 1 1\n");
- fprintf(stdout, "DIMENSION sockets '' absolute 1 1\n");
- fprintf(stdout, "DIMENSION pipes '' absolute 1 1\n");
- fprintf(stdout, "DIMENSION inotifies '' absolute 1 1\n");
- fprintf(stdout, "DIMENSION event '' absolute 1 1\n");
- fprintf(stdout, "DIMENSION timer '' absolute 1 1\n");
- fprintf(stdout, "DIMENSION signal '' absolute 1 1\n");
- fprintf(stdout, "DIMENSION eventpolls '' absolute 1 1\n");
- fprintf(stdout, "DIMENSION other '' absolute 1 1\n");
- }
-
-#ifndef __FreeBSD__
- fprintf(stdout, "CHART %s.%s_uptime '' '%s uptime' 'seconds' uptime %s.uptime line 20250 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION uptime '' absolute 1 1\n");
-
- if (enable_detailed_uptime_charts) {
- fprintf(stdout, "CHART %s.%s_uptime_summary '' '%s uptime summary' 'seconds' uptime %s.uptime_summary area 20255 %d\n", type, w->clean_name, title, type, update_every);
- fprintf(stdout, "CLABEL '%s' '%s' 1\n", lbl_name, w->name);
- fprintf(stdout, "CLABEL_COMMIT\n");
- fprintf(stdout, "DIMENSION min '' absolute 1 1\n");
- fprintf(stdout, "DIMENSION avg '' absolute 1 1\n");
- fprintf(stdout, "DIMENSION max '' absolute 1 1\n");
- }
-#endif
- }
-}
-
-#ifndef __FreeBSD__
-static void send_proc_states_count(usec_t dt)
-{
- static bool chart_added = false;
- // create chart for count of processes in different states
- if (!chart_added) {
- fprintf(
- stdout,
- "CHART system.processes_state '' 'System Processes State' 'processes' processes system.processes_state line %d %d\n",
- NETDATA_CHART_PRIO_SYSTEM_PROCESS_STATES,
- update_every);
- for (proc_state i = PROC_STATUS_RUNNING; i < PROC_STATUS_END; i++) {
- fprintf(stdout, "DIMENSION %s '' absolute 1 1\n", proc_states[i]);
- }
- chart_added = true;
- }
-
- // send process state count
- fprintf(stdout, "BEGIN system.processes_state %" PRIu64 "\n", dt);
- for (proc_state i = PROC_STATUS_RUNNING; i < PROC_STATUS_END; i++) {
- send_SET(proc_states[i], proc_state_count[i]);
- }
- send_END();
-}
-#endif
-
-// ----------------------------------------------------------------------------
-// parse command line arguments
-
-int check_proc_1_io() {
- int ret = 0;
-
- procfile *ff = procfile_open("/proc/1/io", NULL, PROCFILE_FLAG_NO_ERROR_ON_FILE_IO);
- if(!ff) goto cleanup;
-
- ff = procfile_readall(ff);
- if(!ff) goto cleanup;
-
- ret = 1;
-
-cleanup:
- procfile_close(ff);
- return ret;
-}
-
-static void parse_args(int argc, char **argv)
-{
- int i, freq = 0;
-
- for(i = 1; i < argc; i++) {
- if(!freq) {
- int n = (int)str2l(argv[i]);
- if(n > 0) {
- freq = n;
- continue;
- }
- }
-
- if(strcmp("version", argv[i]) == 0 || strcmp("-version", argv[i]) == 0 || strcmp("--version", argv[i]) == 0 || strcmp("-v", argv[i]) == 0 || strcmp("-V", argv[i]) == 0) {
- printf("apps.plugin %s\n", VERSION);
- exit(0);
- }
-
- if(strcmp("test-permissions", argv[i]) == 0 || strcmp("-t", argv[i]) == 0) {
- if(!check_proc_1_io()) {
- perror("Tried to read /proc/1/io and it failed");
- exit(1);
- }
- printf("OK\n");
- exit(0);
- }
-
- if(strcmp("debug", argv[i]) == 0) {
- debug_enabled = 1;
-#ifndef NETDATA_INTERNAL_CHECKS
- fprintf(stderr, "apps.plugin has been compiled without debugging\n");
-#endif
- continue;
- }
-
-#ifndef __FreeBSD__
- if(strcmp("fds-cache-secs", argv[i]) == 0) {
- if(argc <= i + 1) {
- fprintf(stderr, "Parameter 'fds-cache-secs' requires a number as argument.\n");
- exit(1);
- }
- i++;
- max_fds_cache_seconds = str2i(argv[i]);
- if(max_fds_cache_seconds < 0) max_fds_cache_seconds = 0;
- continue;
- }
-#endif
-
- if(strcmp("no-childs", argv[i]) == 0 || strcmp("without-childs", argv[i]) == 0) {
- include_exited_childs = 0;
- continue;
- }
-
- if(strcmp("with-childs", argv[i]) == 0) {
- include_exited_childs = 1;
- continue;
- }
-
- if(strcmp("with-guest", argv[i]) == 0) {
- enable_guest_charts = 1;
- continue;
- }
-
- if(strcmp("no-guest", argv[i]) == 0 || strcmp("without-guest", argv[i]) == 0) {
- enable_guest_charts = 0;
- continue;
- }
-
- if(strcmp("with-files", argv[i]) == 0) {
- enable_file_charts = 1;
- continue;
- }
-
- if(strcmp("no-files", argv[i]) == 0 || strcmp("without-files", argv[i]) == 0) {
- enable_file_charts = 0;
- continue;
- }
-
- if(strcmp("no-users", argv[i]) == 0 || strcmp("without-users", argv[i]) == 0) {
- enable_users_charts = 0;
- continue;
- }
-
- if(strcmp("no-groups", argv[i]) == 0 || strcmp("without-groups", argv[i]) == 0) {
- enable_groups_charts = 0;
- continue;
- }
-
- if(strcmp("with-detailed-uptime", argv[i]) == 0) {
- enable_detailed_uptime_charts = 1;
- continue;
- }
- if(strcmp("with-function-cmdline", argv[i]) == 0) {
- enable_function_cmdline = 1;
- continue;
- }
-
- if(strcmp("-h", argv[i]) == 0 || strcmp("--help", argv[i]) == 0) {
- fprintf(stderr,
- "\n"
- " netdata apps.plugin %s\n"
- " Copyright (C) 2016-2017 Costa Tsaousis <costa@tsaousis.gr>\n"
- " Released under GNU General Public License v3 or later.\n"
- " All rights reserved.\n"
- "\n"
- " This program is a data collector plugin for netdata.\n"
- "\n"
- " Available command line options:\n"
- "\n"
- " SECONDS set the data collection frequency\n"
- "\n"
- " debug enable debugging (lot of output)\n"
- "\n"
- " with-function-cmdline enable reporting the complete command line for processes\n"
- " it includes the command and passed arguments\n"
- " it may include sensitive data such as passwords and tokens\n"
- " enabling this could be a security risk\n"
- "\n"
- " with-childs\n"
- " without-childs enable / disable aggregating exited\n"
- " children resources into parents\n"
- " (default is enabled)\n"
- "\n"
- " with-guest\n"
- " without-guest enable / disable reporting guest charts\n"
- " (default is disabled)\n"
- "\n"
- " with-files\n"
- " without-files enable / disable reporting files, sockets, pipes\n"
- " (default is enabled)\n"
- "\n"
- " without-users disable reporting per user charts\n"
- "\n"
- " without-groups disable reporting per user group charts\n"
- "\n"
- " with-detailed-uptime enable reporting min/avg/max uptime charts\n"
- "\n"
-#ifndef __FreeBSD__
- " fds-cache-secs N cache the files of processed for N seconds\n"
- " caching is adaptive per file (when a file\n"
- " is found, it starts at 0 and while the file\n"
- " remains open, it is incremented up to the\n"
- " max given)\n"
- " (default is %d seconds)\n"
- "\n"
-#endif
- " version or -v or -V print program version and exit\n"
- "\n"
- , VERSION
-#ifndef __FreeBSD__
- , max_fds_cache_seconds
-#endif
- );
- exit(1);
- }
-
- netdata_log_error("Cannot understand option %s", argv[i]);
- exit(1);
- }
-
- if(freq > 0) update_every = freq;
-
- if(read_apps_groups_conf(user_config_dir, "groups")) {
- netdata_log_info("Cannot read process groups configuration file '%s/apps_groups.conf'. Will try '%s/apps_groups.conf'", user_config_dir, stock_config_dir);
-
- if(read_apps_groups_conf(stock_config_dir, "groups")) {
- netdata_log_error("Cannot read process groups '%s/apps_groups.conf'. There are no internal defaults. Failing.", stock_config_dir);
- exit(1);
- }
- else
- netdata_log_info("Loaded config file '%s/apps_groups.conf'", stock_config_dir);
- }
- else
- netdata_log_info("Loaded config file '%s/apps_groups.conf'", user_config_dir);
-}
-
-static int am_i_running_as_root() {
- uid_t uid = getuid(), euid = geteuid();
-
- if(uid == 0 || euid == 0) {
- if(debug_enabled) netdata_log_info("I am running with escalated privileges, uid = %u, euid = %u.", uid, euid);
- return 1;
- }
-
- if(debug_enabled) netdata_log_info("I am not running with escalated privileges, uid = %u, euid = %u.", uid, euid);
- return 0;
-}
-
-#ifdef HAVE_CAPABILITY
-static int check_capabilities() {
- cap_t caps = cap_get_proc();
- if(!caps) {
- netdata_log_error("Cannot get current capabilities.");
- return 0;
- }
- else if(debug_enabled)
- netdata_log_info("Received my capabilities from the system.");
-
- int ret = 1;
-
- cap_flag_value_t cfv = CAP_CLEAR;
- if(cap_get_flag(caps, CAP_DAC_READ_SEARCH, CAP_EFFECTIVE, &cfv) == -1) {
- netdata_log_error("Cannot find if CAP_DAC_READ_SEARCH is effective.");
- ret = 0;
- }
- else {
- if(cfv != CAP_SET) {
- netdata_log_error("apps.plugin should run with CAP_DAC_READ_SEARCH.");
- ret = 0;
- }
- else if(debug_enabled)
- netdata_log_info("apps.plugin runs with CAP_DAC_READ_SEARCH.");
- }
-
- cfv = CAP_CLEAR;
- if(cap_get_flag(caps, CAP_SYS_PTRACE, CAP_EFFECTIVE, &cfv) == -1) {
- netdata_log_error("Cannot find if CAP_SYS_PTRACE is effective.");
- ret = 0;
- }
- else {
- if(cfv != CAP_SET) {
- netdata_log_error("apps.plugin should run with CAP_SYS_PTRACE.");
- ret = 0;
- }
- else if(debug_enabled)
- netdata_log_info("apps.plugin runs with CAP_SYS_PTRACE.");
- }
-
- cap_free(caps);
-
- return ret;
-}
-#else
-static int check_capabilities() {
- return 0;
-}
-#endif
-
-static netdata_mutex_t apps_and_stdout_mutex = NETDATA_MUTEX_INITIALIZER;
-
-#define PROCESS_FILTER_CATEGORY "category:"
-#define PROCESS_FILTER_USER "user:"
-#define PROCESS_FILTER_GROUP "group:"
-#define PROCESS_FILTER_PROCESS "process:"
-#define PROCESS_FILTER_PID "pid:"
-#define PROCESS_FILTER_UID "uid:"
-#define PROCESS_FILTER_GID "gid:"
-
-static struct target *find_target_by_name(struct target *base, const char *name) {
- struct target *t;
- for(t = base; t ; t = t->next) {
- if (strcmp(t->name, name) == 0)
- return t;
- }
-
- return NULL;
-}
-
-static kernel_uint_t MemTotal = 0;
-
-static void get_MemTotal(void) {
-#ifdef __FreeBSD__
- // TODO - fix this for FreeBSD
- return;
-#else
- char filename[FILENAME_MAX + 1];
- snprintfz(filename, FILENAME_MAX, "%s/proc/meminfo", netdata_configured_host_prefix);
-
- procfile *ff = procfile_open(filename, ": \t", PROCFILE_FLAG_DEFAULT);
- if(!ff)
- return;
-
- ff = procfile_readall(ff);
- if(!ff)
- return;
-
- size_t line, lines = procfile_lines(ff);
-
- for(line = 0; line < lines ;line++) {
- size_t words = procfile_linewords(ff, line);
- if(words == 3 && strcmp(procfile_lineword(ff, line, 0), "MemTotal") == 0 && strcmp(procfile_lineword(ff, line, 2), "kB") == 0) {
- kernel_uint_t n = str2ull(procfile_lineword(ff, line, 1), NULL);
- if(n) MemTotal = n;
- break;
- }
- }
-
- procfile_close(ff);
-#endif
-}
-
-static void apps_plugin_function_processes_help(const char *transaction) {
- BUFFER *wb = buffer_create(0, NULL);
- buffer_sprintf(wb, "%s",
- "apps.plugin / processes\n"
- "\n"
- "Function `processes` presents all the currently running processes of the system.\n"
- "\n"
- "The following filters are supported:\n"
- "\n"
- " category:NAME\n"
- " Shows only processes that are assigned the category `NAME` in apps_groups.conf\n"
- "\n"
- " user:NAME\n"
- " Shows only processes that are running as user name `NAME`.\n"
- "\n"
- " group:NAME\n"
- " Shows only processes that are running as group name `NAME`.\n"
- "\n"
- " process:NAME\n"
- " Shows only processes that their Command is `NAME` or their parent's Command is `NAME`.\n"
- "\n"
- " pid:NUMBER\n"
- " Shows only processes that their PID is `NUMBER` or their parent's PID is `NUMBER`\n"
- "\n"
- " uid:NUMBER\n"
- " Shows only processes that their UID is `NUMBER`\n"
- "\n"
- " gid:NUMBER\n"
- " Shows only processes that their GID is `NUMBER`\n"
- "\n"
- "Filters can be combined. Each filter can be given only one time.\n"
- );
-
- pluginsd_function_result_to_stdout(transaction, HTTP_RESP_OK, "text/plain", now_realtime_sec() + 3600, wb);
- buffer_free(wb);
-}
-
-#define add_value_field_llu_with_max(wb, key, value) do { \
- unsigned long long _tmp = (value); \
- key ## _max = (rows == 0) ? (_tmp) : MAX(key ## _max, _tmp); \
- buffer_json_add_array_item_uint64(wb, _tmp); \
-} while(0)
-
-#define add_value_field_ndd_with_max(wb, key, value) do { \
- NETDATA_DOUBLE _tmp = (value); \
- key ## _max = (rows == 0) ? (_tmp) : MAX(key ## _max, _tmp); \
- buffer_json_add_array_item_double(wb, _tmp); \
-} while(0)
-
-static void function_processes(const char *transaction, char *function __maybe_unused, int timeout __maybe_unused, bool *cancelled __maybe_unused) {
- struct pid_stat *p;
-
- char *words[PLUGINSD_MAX_WORDS] = { NULL };
- size_t num_words = quoted_strings_splitter_pluginsd(function, words, PLUGINSD_MAX_WORDS);
-
- struct target *category = NULL, *user = NULL, *group = NULL;
- const char *process_name = NULL;
- pid_t pid = 0;
- uid_t uid = 0;
- gid_t gid = 0;
-
- bool filter_pid = false, filter_uid = false, filter_gid = false;
-
- for(int i = 1; i < PLUGINSD_MAX_WORDS ;i++) {
- const char *keyword = get_word(words, num_words, i);
- if(!keyword) break;
-
- if(!category && strncmp(keyword, PROCESS_FILTER_CATEGORY, strlen(PROCESS_FILTER_CATEGORY)) == 0) {
- category = find_target_by_name(apps_groups_root_target, &keyword[strlen(PROCESS_FILTER_CATEGORY)]);
- if(!category) {
- pluginsd_function_json_error_to_stdout(transaction, HTTP_RESP_BAD_REQUEST,
- "No category with that name found.");
- return;
- }
- }
- else if(!user && strncmp(keyword, PROCESS_FILTER_USER, strlen(PROCESS_FILTER_USER)) == 0) {
- user = find_target_by_name(users_root_target, &keyword[strlen(PROCESS_FILTER_USER)]);
- if(!user) {
- pluginsd_function_json_error_to_stdout(transaction, HTTP_RESP_BAD_REQUEST,
- "No user with that name found.");
- return;
- }
- }
- else if(strncmp(keyword, PROCESS_FILTER_GROUP, strlen(PROCESS_FILTER_GROUP)) == 0) {
- group = find_target_by_name(groups_root_target, &keyword[strlen(PROCESS_FILTER_GROUP)]);
- if(!group) {
- pluginsd_function_json_error_to_stdout(transaction, HTTP_RESP_BAD_REQUEST,
- "No group with that name found.");
- return;
- }
- }
- else if(!process_name && strncmp(keyword, PROCESS_FILTER_PROCESS, strlen(PROCESS_FILTER_PROCESS)) == 0) {
- process_name = &keyword[strlen(PROCESS_FILTER_PROCESS)];
- }
- else if(!pid && strncmp(keyword, PROCESS_FILTER_PID, strlen(PROCESS_FILTER_PID)) == 0) {
- pid = str2i(&keyword[strlen(PROCESS_FILTER_PID)]);
- filter_pid = true;
- }
- else if(!uid && strncmp(keyword, PROCESS_FILTER_UID, strlen(PROCESS_FILTER_UID)) == 0) {
- uid = str2i(&keyword[strlen(PROCESS_FILTER_UID)]);
- filter_uid = true;
- }
- else if(!gid && strncmp(keyword, PROCESS_FILTER_GID, strlen(PROCESS_FILTER_GID)) == 0) {
- gid = str2i(&keyword[strlen(PROCESS_FILTER_GID)]);
- filter_gid = true;
- }
- else if(strcmp(keyword, "help") == 0) {
- apps_plugin_function_processes_help(transaction);
- return;
- }
- else {
- char msg[PLUGINSD_LINE_MAX];
- snprintfz(msg, PLUGINSD_LINE_MAX, "Invalid parameter '%s'", keyword);
- pluginsd_function_json_error_to_stdout(transaction, HTTP_RESP_BAD_REQUEST, msg);
- return;
- }
- }
-
- time_t expires = now_realtime_sec() + update_every;
-
- unsigned int cpu_divisor = time_factor * RATES_DETAIL / 100;
- unsigned int memory_divisor = 1024;
- unsigned int io_divisor = 1024 * RATES_DETAIL;
-
- BUFFER *wb = buffer_create(PLUGINSD_LINE_MAX, NULL);
- buffer_json_initialize(wb, "\"", "\"", 0, true, BUFFER_JSON_OPTIONS_NEWLINE_ON_ARRAY_ITEMS);
- buffer_json_member_add_uint64(wb, "status", HTTP_RESP_OK);
- buffer_json_member_add_string(wb, "type", "table");
- buffer_json_member_add_time_t(wb, "update_every", update_every);
- buffer_json_member_add_string(wb, "help", APPS_PLUGIN_PROCESSES_FUNCTION_DESCRIPTION);
- buffer_json_member_add_array(wb, "data");
-
- NETDATA_DOUBLE
- UserCPU_max = 0.0
- , SysCPU_max = 0.0
- , GuestCPU_max = 0.0
- , CUserCPU_max = 0.0
- , CSysCPU_max = 0.0
- , CGuestCPU_max = 0.0
- , CPU_max = 0.0
- , VMSize_max = 0.0
- , RSS_max = 0.0
- , Shared_max = 0.0
- , Swap_max = 0.0
- , Memory_max = 0.0
- , FDsLimitPercent_max = 0.0
- ;
-
- unsigned long long
- Processes_max = 0
- , Threads_max = 0
- , VoluntaryCtxtSwitches_max = 0
- , NonVoluntaryCtxtSwitches_max = 0
- , Uptime_max = 0
- , MinFlt_max = 0
- , CMinFlt_max = 0
- , TMinFlt_max = 0
- , MajFlt_max = 0
- , CMajFlt_max = 0
- , TMajFlt_max = 0
- , PReads_max = 0
- , PWrites_max = 0
- , RCalls_max = 0
- , WCalls_max = 0
- , Files_max = 0
- , Pipes_max = 0
- , Sockets_max = 0
- , iNotiFDs_max = 0
- , EventFDs_max = 0
- , TimerFDs_max = 0
- , SigFDs_max = 0
- , EvPollFDs_max = 0
- , OtherFDs_max = 0
- , FDs_max = 0
- ;
-
-#ifndef __FreeBSD__
- unsigned long long
- LReads_max = 0
- , LWrites_max = 0
- ;
-#endif
-
- int rows= 0;
- for(p = root_of_pids; p ; p = p->next) {
- if(!p->updated)
- continue;
-
- if(category && p->target != category)
- continue;
-
- if(user && p->user_target != user)
- continue;
-
- if(group && p->group_target != group)
- continue;
-
- if(process_name && ((strcmp(p->comm, process_name) != 0 && !p->parent) || (p->parent && strcmp(p->comm, process_name) != 0 && strcmp(p->parent->comm, process_name) != 0)))
- continue;
-
- if(filter_pid && p->pid != pid && p->ppid != pid)
- continue;
-
- if(filter_uid && p->uid != uid)
- continue;
-
- if(filter_gid && p->gid != gid)
- continue;
-
- rows++;
-
- buffer_json_add_array_item_array(wb); // for each pid
-
- // IMPORTANT!
- // THE ORDER SHOULD BE THE SAME WITH THE FIELDS!
-
- // pid
- buffer_json_add_array_item_uint64(wb, p->pid);
-
- // cmd
- buffer_json_add_array_item_string(wb, p->comm);
-
- // cmdline
- if (enable_function_cmdline) {
- buffer_json_add_array_item_string(wb, (p->cmdline && *p->cmdline) ? p->cmdline : p->comm);
- }
-
- // ppid
- buffer_json_add_array_item_uint64(wb, p->ppid);
-
- // category
- buffer_json_add_array_item_string(wb, p->target ? p->target->name : "-");
-
- // user
- buffer_json_add_array_item_string(wb, p->user_target ? p->user_target->name : "-");
-
- // uid
- buffer_json_add_array_item_uint64(wb, p->uid);
-
- // group
- buffer_json_add_array_item_string(wb, p->group_target ? p->group_target->name : "-");
-
- // gid
- buffer_json_add_array_item_uint64(wb, p->gid);
-
- // CPU utilization %
- add_value_field_ndd_with_max(wb, CPU, (NETDATA_DOUBLE)(p->utime + p->stime + p->gtime + p->cutime + p->cstime + p->cgtime) / cpu_divisor);
- add_value_field_ndd_with_max(wb, UserCPU, (NETDATA_DOUBLE)(p->utime) / cpu_divisor);
- add_value_field_ndd_with_max(wb, SysCPU, (NETDATA_DOUBLE)(p->stime) / cpu_divisor);
- add_value_field_ndd_with_max(wb, GuestCPU, (NETDATA_DOUBLE)(p->gtime) / cpu_divisor);
- add_value_field_ndd_with_max(wb, CUserCPU, (NETDATA_DOUBLE)(p->cutime) / cpu_divisor);
- add_value_field_ndd_with_max(wb, CSysCPU, (NETDATA_DOUBLE)(p->cstime) / cpu_divisor);
- add_value_field_ndd_with_max(wb, CGuestCPU, (NETDATA_DOUBLE)(p->cgtime) / cpu_divisor);
-
- add_value_field_llu_with_max(wb, VoluntaryCtxtSwitches, p->status_voluntary_ctxt_switches / RATES_DETAIL);
- add_value_field_llu_with_max(wb, NonVoluntaryCtxtSwitches, p->status_nonvoluntary_ctxt_switches / RATES_DETAIL);
-
- // memory MiB
- if(MemTotal)
- add_value_field_ndd_with_max(wb, Memory, (NETDATA_DOUBLE)p->status_vmrss * 100.0 / (NETDATA_DOUBLE)MemTotal);
-
- add_value_field_ndd_with_max(wb, RSS, (NETDATA_DOUBLE)p->status_vmrss / memory_divisor);
- add_value_field_ndd_with_max(wb, Shared, (NETDATA_DOUBLE)p->status_vmshared / memory_divisor);
- add_value_field_ndd_with_max(wb, VMSize, (NETDATA_DOUBLE)p->status_vmsize / memory_divisor);
- add_value_field_ndd_with_max(wb, Swap, (NETDATA_DOUBLE)p->status_vmswap / memory_divisor);
-
- // Physical I/O
- add_value_field_llu_with_max(wb, PReads, p->io_storage_bytes_read / io_divisor);
- add_value_field_llu_with_max(wb, PWrites, p->io_storage_bytes_written / io_divisor);
-
- // Logical I/O
-#ifndef __FreeBSD__
- add_value_field_llu_with_max(wb, LReads, p->io_logical_bytes_read / io_divisor);
- add_value_field_llu_with_max(wb, LWrites, p->io_logical_bytes_written / io_divisor);
-#endif
-
- // I/O calls
- add_value_field_llu_with_max(wb, RCalls, p->io_read_calls / RATES_DETAIL);
- add_value_field_llu_with_max(wb, WCalls, p->io_write_calls / RATES_DETAIL);
-
- // minor page faults
- add_value_field_llu_with_max(wb, MinFlt, p->minflt / RATES_DETAIL);
- add_value_field_llu_with_max(wb, CMinFlt, p->cminflt / RATES_DETAIL);
- add_value_field_llu_with_max(wb, TMinFlt, (p->minflt + p->cminflt) / RATES_DETAIL);
-
- // major page faults
- add_value_field_llu_with_max(wb, MajFlt, p->majflt / RATES_DETAIL);
- add_value_field_llu_with_max(wb, CMajFlt, p->cmajflt / RATES_DETAIL);
- add_value_field_llu_with_max(wb, TMajFlt, (p->majflt + p->cmajflt) / RATES_DETAIL);
-
- // open file descriptors
- add_value_field_ndd_with_max(wb, FDsLimitPercent, p->openfds_limits_percent);
- add_value_field_llu_with_max(wb, FDs, pid_openfds_sum(p));
- add_value_field_llu_with_max(wb, Files, p->openfds.files);
- add_value_field_llu_with_max(wb, Pipes, p->openfds.pipes);
- add_value_field_llu_with_max(wb, Sockets, p->openfds.sockets);
- add_value_field_llu_with_max(wb, iNotiFDs, p->openfds.inotifies);
- add_value_field_llu_with_max(wb, EventFDs, p->openfds.eventfds);
- add_value_field_llu_with_max(wb, TimerFDs, p->openfds.timerfds);
- add_value_field_llu_with_max(wb, SigFDs, p->openfds.signalfds);
- add_value_field_llu_with_max(wb, EvPollFDs, p->openfds.eventpolls);
- add_value_field_llu_with_max(wb, OtherFDs, p->openfds.other);
-
-
- // processes, threads, uptime
- add_value_field_llu_with_max(wb, Processes, p->children_count);
- add_value_field_llu_with_max(wb, Threads, p->num_threads);
- add_value_field_llu_with_max(wb, Uptime, p->uptime);
-
- buffer_json_array_close(wb); // for each pid
- }
-
- buffer_json_array_close(wb); // data
- buffer_json_member_add_object(wb, "columns");
-
- {
- int field_id = 0;
-
- // IMPORTANT!
- // THE ORDER SHOULD BE THE SAME WITH THE VALUES!
- // wb, key, name, visible, type, visualization, transform, decimal_points, units, max, sort, sortable, sticky, unique_key, pointer_to, summary, range
- buffer_rrdf_table_add_field(wb, field_id++, "PID", "Process ID", RRDF_FIELD_TYPE_INTEGER,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NUMBER, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY |
- RRDF_FIELD_OPTS_UNIQUE_KEY, NULL);
-
- buffer_rrdf_table_add_field(wb, field_id++, "Cmd", "Process Name", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
-
- if (enable_function_cmdline) {
- buffer_rrdf_table_add_field(wb, field_id++, "CmdLine", "Command Line", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0,
- NULL, NAN, RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_NONE, NULL);
- }
-
- buffer_rrdf_table_add_field(wb, field_id++, "PPID", "Parent Process ID", RRDF_FIELD_TYPE_INTEGER,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NUMBER, 0, NULL,
- NAN, RRDF_FIELD_SORT_ASCENDING, "PID", RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Category", "Category (apps_groups.conf)", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE,
- RRDF_FIELD_TRANSFORM_NONE,
- 0, NULL, NAN, RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "User", "User Owner", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Uid", "User ID", RRDF_FIELD_TYPE_INTEGER, RRDF_FIELD_VISUAL_VALUE,
- RRDF_FIELD_TRANSFORM_NUMBER, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Group", "Group Owner", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Gid", "Group ID", RRDF_FIELD_TYPE_INTEGER, RRDF_FIELD_VISUAL_VALUE,
- RRDF_FIELD_TRANSFORM_NUMBER, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_NONE, NULL);
-
- // CPU utilization
- buffer_rrdf_table_add_field(wb, field_id++, "CPU", "Total CPU Time (100% = 1 core)",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "%", CPU_max, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_VISIBLE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "UserCPU", "User CPU time (100% = 1 core)",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 2, "%", UserCPU_max,
- RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "SysCPU", "System CPU Time (100% = 1 core)",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 2, "%", SysCPU_max,
- RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "GuestCPU", "Guest CPU Time (100% = 1 core)",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 2, "%", GuestCPU_max,
- RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "CUserCPU", "Children User CPU Time (100% = 1 core)",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "%", CUserCPU_max, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "CSysCPU", "Children System CPU Time (100% = 1 core)",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "%", CSysCPU_max, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "CGuestCPU", "Children Guest CPU Time (100% = 1 core)",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "%", CGuestCPU_max, RRDF_FIELD_SORT_DESCENDING,
- NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE, RRDF_FIELD_OPTS_NONE, NULL);
-
- // CPU context switches
- buffer_rrdf_table_add_field(wb, field_id++, "vCtxSwitch", "Voluntary Context Switches",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 2, "switches/s",
- VoluntaryCtxtSwitches_max, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE, RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "iCtxSwitch", "Involuntary Context Switches",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 2, "switches/s",
- NonVoluntaryCtxtSwitches_max, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE, RRDF_FIELD_OPTS_NONE, NULL);
-
- // memory
- if (MemTotal)
- buffer_rrdf_table_add_field(wb, field_id++, "Memory", "Memory Percentage", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "%", 100.0, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_VISIBLE, NULL);
-
- buffer_rrdf_table_add_field(wb, field_id++, "Resident", "Resident Set Size", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER,
- 2, "MiB", RSS_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_VISIBLE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Shared", "Shared Pages", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 2,
- "MiB", Shared_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_VISIBLE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Virtual", "Virtual Memory Size", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "MiB", VMSize_max, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_VISIBLE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Swap", "Swap Memory", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 2,
- "MiB",
- Swap_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
-
- // Physical I/O
- buffer_rrdf_table_add_field(wb, field_id++, "PReads", "Physical I/O Reads", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER,
- 2, "KiB/s", PReads_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_VISIBLE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "PWrites", "Physical I/O Writes", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "KiB/s", PWrites_max, RRDF_FIELD_SORT_DESCENDING,
- NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_VISIBLE, NULL);
-
- // Logical I/O
-#ifndef __FreeBSD__
- buffer_rrdf_table_add_field(wb, field_id++, "LReads", "Logical I/O Reads", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER,
- 2, "KiB/s", LReads_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "LWrites", "Logical I/O Writes", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER,
- 2, "KiB/s", LWrites_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
-#endif
-
- // I/O calls
- buffer_rrdf_table_add_field(wb, field_id++, "RCalls", "I/O Read Calls", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 2,
- "calls/s", RCalls_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "WCalls", "I/O Write Calls", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 2,
- "calls/s", WCalls_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
-
- // minor page faults
- buffer_rrdf_table_add_field(wb, field_id++, "MinFlt", "Minor Page Faults/s", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER,
- 2, "pgflts/s", MinFlt_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "CMinFlt", "Children Minor Page Faults/s",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "pgflts/s", CMinFlt_max, RRDF_FIELD_SORT_DESCENDING,
- NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "TMinFlt", "Total Minor Page Faults/s",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "pgflts/s", TMinFlt_max, RRDF_FIELD_SORT_DESCENDING,
- NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
-
- // major page faults
- buffer_rrdf_table_add_field(wb, field_id++, "MajFlt", "Major Page Faults/s", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER,
- 2, "pgflts/s", MajFlt_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "CMajFlt", "Children Major Page Faults/s",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "pgflts/s", CMajFlt_max, RRDF_FIELD_SORT_DESCENDING,
- NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "TMajFlt", "Total Major Page Faults/s",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "pgflts/s", TMajFlt_max, RRDF_FIELD_SORT_DESCENDING,
- NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
-
- // open file descriptors
- buffer_rrdf_table_add_field(wb, field_id++, "FDsLimitPercent", "Percentage of Open Descriptors vs Limits",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 2, "%", FDsLimitPercent_max, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_MAX, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "FDs", "All Open File Descriptors",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 0, "fds", FDs_max, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Files", "Open Files", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 0,
- "fds",
- Files_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Pipes", "Open Pipes", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 0,
- "fds",
- Pipes_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Sockets", "Open Sockets", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 0,
- "fds", Sockets_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "iNotiFDs", "Open iNotify Descriptors",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 0, "fds", iNotiFDs_max, RRDF_FIELD_SORT_DESCENDING,
- NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "EventFDs", "Open Event Descriptors",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 0, "fds", EventFDs_max, RRDF_FIELD_SORT_DESCENDING,
- NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "TimerFDs", "Open Timer Descriptors",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 0, "fds", TimerFDs_max, RRDF_FIELD_SORT_DESCENDING,
- NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "SigFDs", "Open Signal Descriptors",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 0, "fds", SigFDs_max, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "EvPollFDs", "Open Event Poll Descriptors",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 0, "fds", EvPollFDs_max,
- RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "OtherFDs", "Other Open Descriptors",
- RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR,
- RRDF_FIELD_TRANSFORM_NUMBER, 0, "fds", OtherFDs_max, RRDF_FIELD_SORT_DESCENDING,
- NULL, RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
-
- // processes, threads, uptime
- buffer_rrdf_table_add_field(wb, field_id++, "Processes", "Processes", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 0,
- "processes", Processes_max, RRDF_FIELD_SORT_DESCENDING, NULL,
- RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Threads", "Threads", RRDF_FIELD_TYPE_BAR_WITH_INTEGER,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER, 0,
- "threads", Threads_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_SUM,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_NONE, NULL);
- buffer_rrdf_table_add_field(wb, field_id++, "Uptime", "Uptime in seconds", RRDF_FIELD_TYPE_DURATION,
- RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_DURATION_S, 2,
- "seconds", Uptime_max, RRDF_FIELD_SORT_DESCENDING, NULL, RRDF_FIELD_SUMMARY_MAX,
- RRDF_FIELD_FILTER_RANGE,
- RRDF_FIELD_OPTS_VISIBLE, NULL);
- }
- buffer_json_object_close(wb); // columns
-
- buffer_json_member_add_string(wb, "default_sort_column", "CPU");
-
- buffer_json_member_add_object(wb, "charts");
- {
- // CPU chart
- buffer_json_member_add_object(wb, "CPU");
- {
- buffer_json_member_add_string(wb, "name", "CPU Utilization");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "UserCPU");
- buffer_json_add_array_item_string(wb, "SysCPU");
- buffer_json_add_array_item_string(wb, "GuestCPU");
- buffer_json_add_array_item_string(wb, "CUserCPU");
- buffer_json_add_array_item_string(wb, "CSysCPU");
- buffer_json_add_array_item_string(wb, "CGuestCPU");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- buffer_json_member_add_object(wb, "CPUCtxSwitches");
- {
- buffer_json_member_add_string(wb, "name", "CPU Context Switches");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "vCtxSwitch");
- buffer_json_add_array_item_string(wb, "iCtxSwitch");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // Memory chart
- buffer_json_member_add_object(wb, "Memory");
- {
- buffer_json_member_add_string(wb, "name", "Memory");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Virtual");
- buffer_json_add_array_item_string(wb, "Resident");
- buffer_json_add_array_item_string(wb, "Shared");
- buffer_json_add_array_item_string(wb, "Swap");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- if(MemTotal) {
- // Memory chart
- buffer_json_member_add_object(wb, "MemoryPercent");
- {
- buffer_json_member_add_string(wb, "name", "Memory Percentage");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Memory");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
- }
-
-#ifndef __FreeBSD__
- // I/O Reads chart
- buffer_json_member_add_object(wb, "Reads");
- {
- buffer_json_member_add_string(wb, "name", "I/O Reads");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "LReads");
- buffer_json_add_array_item_string(wb, "PReads");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // I/O Writes chart
- buffer_json_member_add_object(wb, "Writes");
- {
- buffer_json_member_add_string(wb, "name", "I/O Writes");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "LWrites");
- buffer_json_add_array_item_string(wb, "PWrites");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // Logical I/O chart
- buffer_json_member_add_object(wb, "LogicalIO");
- {
- buffer_json_member_add_string(wb, "name", "Logical I/O");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "LReads");
- buffer_json_add_array_item_string(wb, "LWrites");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-#endif
-
- // Physical I/O chart
- buffer_json_member_add_object(wb, "PhysicalIO");
- {
- buffer_json_member_add_string(wb, "name", "Physical I/O");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "PReads");
- buffer_json_add_array_item_string(wb, "PWrites");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // I/O Calls chart
- buffer_json_member_add_object(wb, "IOCalls");
- {
- buffer_json_member_add_string(wb, "name", "I/O Calls");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "RCalls");
- buffer_json_add_array_item_string(wb, "WCalls");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // Minor Page Faults chart
- buffer_json_member_add_object(wb, "MinFlt");
- {
- buffer_json_member_add_string(wb, "name", "Minor Page Faults");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "MinFlt");
- buffer_json_add_array_item_string(wb, "CMinFlt");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // Major Page Faults chart
- buffer_json_member_add_object(wb, "MajFlt");
- {
- buffer_json_member_add_string(wb, "name", "Major Page Faults");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "MajFlt");
- buffer_json_add_array_item_string(wb, "CMajFlt");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // Threads chart
- buffer_json_member_add_object(wb, "Threads");
- {
- buffer_json_member_add_string(wb, "name", "Threads");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Threads");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // Processes chart
- buffer_json_member_add_object(wb, "Processes");
- {
- buffer_json_member_add_string(wb, "name", "Processes");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Processes");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // FDs chart
- buffer_json_member_add_object(wb, "FDs");
- {
- buffer_json_member_add_string(wb, "name", "File Descriptors");
- buffer_json_member_add_string(wb, "type", "stacked-bar");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Files");
- buffer_json_add_array_item_string(wb, "Pipes");
- buffer_json_add_array_item_string(wb, "Sockets");
- buffer_json_add_array_item_string(wb, "iNotiFDs");
- buffer_json_add_array_item_string(wb, "EventFDs");
- buffer_json_add_array_item_string(wb, "TimerFDs");
- buffer_json_add_array_item_string(wb, "SigFDs");
- buffer_json_add_array_item_string(wb, "EvPollFDs");
- buffer_json_add_array_item_string(wb, "OtherFDs");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
- }
- buffer_json_object_close(wb); // charts
-
- buffer_json_member_add_array(wb, "default_charts");
- {
- buffer_json_add_array_item_array(wb);
- buffer_json_add_array_item_string(wb, "CPU");
- buffer_json_add_array_item_string(wb, "Category");
- buffer_json_array_close(wb);
-
- buffer_json_add_array_item_array(wb);
- buffer_json_add_array_item_string(wb, "Memory");
- buffer_json_add_array_item_string(wb, "Category");
- buffer_json_array_close(wb);
- }
- buffer_json_array_close(wb);
-
- buffer_json_member_add_object(wb, "group_by");
- {
- // group by PID
- buffer_json_member_add_object(wb, "PID");
- {
- buffer_json_member_add_string(wb, "name", "Process Tree by PID");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "PPID");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // group by Category
- buffer_json_member_add_object(wb, "Category");
- {
- buffer_json_member_add_string(wb, "name", "Process Tree by Category");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Category");
- buffer_json_add_array_item_string(wb, "PPID");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // group by User
- buffer_json_member_add_object(wb, "User");
- {
- buffer_json_member_add_string(wb, "name", "Process Tree by User");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "User");
- buffer_json_add_array_item_string(wb, "PPID");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // group by Group
- buffer_json_member_add_object(wb, "Group");
- {
- buffer_json_member_add_string(wb, "name", "Process Tree by Group");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Group");
- buffer_json_add_array_item_string(wb, "PPID");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
- }
- buffer_json_object_close(wb); // group_by
-
- buffer_json_member_add_time_t(wb, "expires", expires);
- buffer_json_finalize(wb);
-
- pluginsd_function_result_to_stdout(transaction, HTTP_RESP_OK, "application/json", expires, wb);
-
- buffer_free(wb);
-}
-
-static bool apps_plugin_exit = false;
-
-int main(int argc, char **argv) {
- clocks_init();
- nd_log_initialize_for_external_plugins("apps.plugin");
-
- pagesize = (size_t)sysconf(_SC_PAGESIZE);
-
- bool send_resource_usage = true;
- {
- const char *s = getenv("NETDATA_INTERNALS_MONITORING");
- if(s && *s && strcmp(s, "NO") == 0)
- send_resource_usage = false;
- }
-
- // since apps.plugin runs as root, prevent it from opening symbolic links
- procfile_open_flags = O_RDONLY|O_NOFOLLOW;
-
- netdata_configured_host_prefix = getenv("NETDATA_HOST_PREFIX");
- if(verify_netdata_host_prefix(true) == -1) exit(1);
-
- user_config_dir = getenv("NETDATA_USER_CONFIG_DIR");
- if(user_config_dir == NULL) {
- // netdata_log_info("NETDATA_CONFIG_DIR is not passed from netdata");
- user_config_dir = CONFIG_DIR;
- }
- // else netdata_log_info("Found NETDATA_USER_CONFIG_DIR='%s'", user_config_dir);
-
- stock_config_dir = getenv("NETDATA_STOCK_CONFIG_DIR");
- if(stock_config_dir == NULL) {
- // netdata_log_info("NETDATA_CONFIG_DIR is not passed from netdata");
- stock_config_dir = LIBCONFIG_DIR;
- }
- // else netdata_log_info("Found NETDATA_USER_CONFIG_DIR='%s'", user_config_dir);
-
-#ifdef NETDATA_INTERNAL_CHECKS
- if(debug_flags != 0) {
- struct rlimit rl = { RLIM_INFINITY, RLIM_INFINITY };
- if(setrlimit(RLIMIT_CORE, &rl) != 0)
- netdata_log_info("Cannot request unlimited core dumps for debugging... Proceeding anyway...");
-#ifdef HAVE_SYS_PRCTL_H
- prctl(PR_SET_DUMPABLE, 1, 0, 0, 0);
-#endif
- }
-#endif /* NETDATA_INTERNAL_CHECKS */
-
- procfile_adaptive_initial_allocation = 1;
-
- get_system_HZ();
-#ifdef __FreeBSD__
- time_factor = 1000000ULL / RATES_DETAIL; // FreeBSD uses usecs
-#else
- time_factor = system_hz; // Linux uses clock ticks
-#endif
-
- get_system_pid_max();
- get_system_cpus_uncached();
-
- parse_args(argc, argv);
-
- if(!check_capabilities() && !am_i_running_as_root() && !check_proc_1_io()) {
- uid_t uid = getuid(), euid = geteuid();
-#ifdef HAVE_CAPABILITY
- netdata_log_error("apps.plugin should either run as root (now running with uid %u, euid %u) or have special capabilities. "
- "Without these, apps.plugin cannot report disk I/O utilization of other processes. "
- "To enable capabilities run: sudo setcap cap_dac_read_search,cap_sys_ptrace+ep %s; "
- "To enable setuid to root run: sudo chown root:netdata %s; sudo chmod 4750 %s; "
- , uid, euid, argv[0], argv[0], argv[0]
- );
-#else
- netdata_log_error("apps.plugin should either run as root (now running with uid %u, euid %u) or have special capabilities. "
- "Without these, apps.plugin cannot report disk I/O utilization of other processes. "
- "Your system does not support capabilities. "
- "To enable setuid to root run: sudo chown root:netdata %s; sudo chmod 4750 %s; "
- , uid, euid, argv[0], argv[0]
- );
-#endif
- }
-
- netdata_log_info("started on pid %d", getpid());
-
- snprintfz(all_user_ids.filename, FILENAME_MAX, "%s/etc/passwd", netdata_configured_host_prefix);
- debug_log("passwd file: '%s'", all_user_ids.filename);
-
- snprintfz(all_group_ids.filename, FILENAME_MAX, "%s/etc/group", netdata_configured_host_prefix);
- debug_log("group file: '%s'", all_group_ids.filename);
-
-#if (ALL_PIDS_ARE_READ_INSTANTLY == 0)
- all_pids_sortlist = callocz(sizeof(pid_t), (size_t)pid_max + 1);
-#endif
-
- all_pids = callocz(sizeof(struct pid_stat *), (size_t) pid_max + 1);
-
- // ------------------------------------------------------------------------
- // the event loop for functions
-
- struct functions_evloop_globals *wg =
- functions_evloop_init(1, "APPS", &apps_and_stdout_mutex, &apps_plugin_exit);
-
- functions_evloop_add_function(wg, "processes", function_processes, PLUGINS_FUNCTIONS_TIMEOUT_DEFAULT);
-
- // ------------------------------------------------------------------------
-
- netdata_mutex_lock(&apps_and_stdout_mutex);
- APPS_PLUGIN_GLOBAL_FUNCTIONS();
-
- usec_t step = update_every * USEC_PER_SEC;
- global_iterations_counter = 1;
- heartbeat_t hb;
- heartbeat_init(&hb);
- for(; !apps_plugin_exit ; global_iterations_counter++) {
- netdata_mutex_unlock(&apps_and_stdout_mutex);
-
-#ifdef NETDATA_PROFILING
-#warning "compiling for profiling"
- static int profiling_count=0;
- profiling_count++;
- if(unlikely(profiling_count > 2000)) exit(0);
- usec_t dt = update_every * USEC_PER_SEC;
-#else
- usec_t dt = heartbeat_next(&hb, step);
-#endif
- netdata_mutex_lock(&apps_and_stdout_mutex);
-
- struct pollfd pollfd = { .fd = fileno(stdout), .events = POLLERR };
- if (unlikely(poll(&pollfd, 1, 0) < 0)) {
- netdata_mutex_unlock(&apps_and_stdout_mutex);
- fatal("Cannot check if a pipe is available");
- }
- if (unlikely(pollfd.revents & POLLERR)) {
- netdata_mutex_unlock(&apps_and_stdout_mutex);
- fatal("Received error on read pipe.");
- }
-
- if(global_iterations_counter % 10 == 0)
- get_MemTotal();
-
- if(!collect_data_for_all_processes()) {
- netdata_log_error("Cannot collect /proc data for running processes. Disabling apps.plugin...");
- printf("DISABLE\n");
- netdata_mutex_unlock(&apps_and_stdout_mutex);
- exit(1);
- }
-
- calculate_netdata_statistics();
- normalize_utilization(apps_groups_root_target);
-
- if(send_resource_usage)
- send_resource_usage_to_netdata(dt);
-
-#ifndef __FreeBSD__
- send_proc_states_count(dt);
-#endif
-
- send_charts_updates_to_netdata(apps_groups_root_target, "app", "app_group", "Apps");
- send_collected_data_to_netdata(apps_groups_root_target, "app", dt);
-
- if (enable_users_charts) {
- send_charts_updates_to_netdata(users_root_target, "user", "user", "Users");
- send_collected_data_to_netdata(users_root_target, "user", dt);
- }
-
- if (enable_groups_charts) {
- send_charts_updates_to_netdata(groups_root_target, "usergroup", "user_group", "User Groups");
- send_collected_data_to_netdata(groups_root_target, "usergroup", dt);
- }
-
- fflush(stdout);
-
- show_guest_time_old = show_guest_time;
-
- debug_log("done Loop No %zu", global_iterations_counter);
- }
- netdata_mutex_unlock(&apps_and_stdout_mutex);
-}
diff --git a/collectors/cgroups.plugin/Makefile.am b/collectors/cgroups.plugin/Makefile.am
deleted file mode 100644
index 0f6062420..000000000
--- a/collectors/cgroups.plugin/Makefile.am
+++ /dev/null
@@ -1,23 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-CLEANFILES = \
- cgroup-name.sh \
- cgroup-network-helper.sh \
- $(NULL)
-
-include $(top_srcdir)/build/subst.inc
-SUFFIXES = .in
-
-dist_plugins_SCRIPTS = \
- cgroup-name.sh \
- cgroup-network-helper.sh \
- $(NULL)
-
-dist_noinst_DATA = \
- cgroup-name.sh.in \
- cgroup-network-helper.sh.in \
- README.md \
- $(NULL)
diff --git a/collectors/cgroups.plugin/README.md b/collectors/cgroups.plugin/README.md
deleted file mode 100644
index ba6a20e5e..000000000
--- a/collectors/cgroups.plugin/README.md
+++ /dev/null
@@ -1,302 +0,0 @@
-<!--
-title: "Monitor Cgroups (cgroups.plugin)"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/README.md"
-sidebar_label: "Monitor Cgroups"
-learn_status: "Published"
-learn_topic_type: "References"
-learn_rel_path: "Integrations/Monitor/Virtualized environments/Containers"
--->
-
-# Monitor Cgroups (cgroups.plugin)
-
-You can monitor containers and virtual machines using **cgroups**.
-
-cgroups (or control groups), are a Linux kernel feature that provides accounting and resource usage limiting for
-processes. When cgroups are bundled with namespaces (i.e. isolation), they form what we usually call **containers**.
-
-cgroups are hierarchical, meaning that cgroups can contain child cgroups, which can contain more cgroups, etc. All
-accounting is reported (and resource usage limits are applied) also in a hierarchical way.
-
-To visualize cgroup metrics Netdata provides configuration for cherry picking the cgroups of interest. By default (
-without any configuration) Netdata should pick **systemd services**, all kinds of **containers** (lxc, docker, etc)
-and **virtual machines** spawn by managers that register them with cgroups (qemu, libvirt, etc).
-
-## Configuring Netdata for cgroups
-
-In general, no additional settings are required. Netdata discovers all available cgroups on the host system and
-collects their metrics.
-
-### How Netdata finds the available cgroups
-
-Linux exposes resource usage reporting and provides dynamic configuration for cgroups, using virtual files (usually)
-under `/sys/fs/cgroup`. Netdata reads `/proc/self/mountinfo` to detect the exact mount point of cgroups. Netdata also
-allows manual configuration of this mount point, using these settings:
-
-```text
-[plugin:cgroups]
- check for new cgroups every = 10
- path to /sys/fs/cgroup/cpuacct = /sys/fs/cgroup/cpuacct
- path to /sys/fs/cgroup/blkio = /sys/fs/cgroup/blkio
- path to /sys/fs/cgroup/memory = /sys/fs/cgroup/memory
- path to /sys/fs/cgroup/devices = /sys/fs/cgroup/devices
-```
-
-Netdata rescans these directories for added or removed cgroups every `check for new cgroups every` seconds.
-
-### Hierarchical search for cgroups
-
-Since cgroups are hierarchical, for each of the directories shown above, Netdata walks through the subdirectories
-recursively searching for cgroups (each subdirectory is another cgroup).
-
-To provide a sane default for this setting, Netdata uses the following pattern list (patterns starting with `!` give a
-negative match and their order is important: the first matching a path will be used):
-
-```text
-[plugin:cgroups]
- search for cgroups in subpaths matching = !*/init.scope !*-qemu !/init.scope !/system !/systemd !/user !/user.slice *
-```
-
-So, we disable checking for **child cgroups** in systemd internal
-cgroups ([systemd services are monitored by Netdata](#monitoring-systemd-services)), user cgroups (normally used for
-desktop and remote user sessions), qemu virtual machines (child cgroups of virtual machines) and `init.scope`. All
-others are enabled.
-
-### Unified cgroups (cgroups v2) support
-
-Netdata automatically detects cgroups version. If detection fails Netdata assumes v1.
-To switch to v2 manually add:
-
-```text
-[plugin:cgroups]
- use unified cgroups = yes
- path to unified cgroups = /sys/fs/cgroup
-```
-
-Unified cgroups use same name pattern matching as v1 cgroups. `cgroup_enable_systemd_services_detailed_memory` is
-currently unsupported when using unified cgroups.
-
-### Enabled cgroups
-
-To provide a sane default, Netdata uses the
-following [pattern list](https://github.com/netdata/netdata/blob/master/libnetdata/simple_pattern/README.md):
-
-- Checks the pattern against the path of the cgroup
-
- ```text
- [plugin:cgroups]
- enable by default cgroups matching = !*/init.scope *.scope !*/vcpu* !*/emulator !*.mount !*.partition !*.service !*.slice !*.swap !*.user !/ !/docker !/libvirt !/lxc !/lxc/*/ns !/lxc/*/ns/* !/machine !/qemu !/system !/systemd !/user *
- ```
-
-- Checks the pattern against the name of the cgroup (as you see it on the dashboard)
-
- ```text
- [plugin:cgroups]
- enable by default cgroups names matching = *
- ```
-
-Renaming is configured with the following options:
-
-```text
-[plugin:cgroups]
- run script to rename cgroups matching = *.scope *docker* *lxc* *qemu* !/ !*.mount !*.partition !*.service !*.slice !*.swap !*.user *
- script to get cgroup names = /usr/libexec/netdata/plugins.d/cgroup-name.sh
-```
-
-The whole point for the additional pattern list, is to limit the number of times the script will be called. Without this
-pattern list, the script might be called thousands of times, depending on the number of cgroups available in the system.
-
-The above pattern list is matched against the path of the cgroup. For matched cgroups, Netdata calls the
-script [cgroup-name.sh](https://raw.githubusercontent.com/netdata/netdata/master/collectors/cgroups.plugin/cgroup-name.sh)
-to get its name. This script queries `docker`, `kubectl`, `podman`, or applies heuristics to find give a name for the
-cgroup.
-
-#### Note on Podman container names
-
-Podman's security model is a lot more restrictive than Docker's, so Netdata will not be able to detect container names
-out of the box unless they were started by the same user as Netdata itself.
-
-If Podman is used in "rootful" mode, it's also possible to use `podman system service` to grant Netdata access to
-container names. To do this, ensure `podman system service` is running and Netdata has access
-to `/run/podman/podman.sock` (the default permissions as specified by upstream are `0600`, with owner `root`, so you
-will have to adjust the configuration).
-
-[Docker Socket Proxy (HAProxy)](https://github.com/Tecnativa/docker-socket-proxy) or [CetusGuard](https://github.com/hectorm/cetusguard)
-can also be used to give Netdata restricted access to the socket. Note that `PODMAN_HOST` in Netdata's environment should
-be set to the proxy's URL in this case.
-
-### Charts with zero metrics
-
-By default, Netdata will enable monitoring metrics only when they are not zero. If they are constantly zero they are
-ignored. Metrics that will start having values, after Netdata is started, will be detected and charts will be
-automatically added to the dashboard (a refresh of the dashboard is needed for them to appear though). Set `yes` for a
-chart instead of `auto` to enable it permanently. For example:
-
-```text
-[plugin:cgroups]
- enable memory (used mem including cache) = yes
-```
-
-You can also set the `enable zero metrics` option to `yes` in the `[global]` section which enables charts with zero
-metrics for all internal Netdata plugins.
-
-### Alerts
-
-CPU and memory limits are watched and used to rise alerts. Memory usage for every cgroup is checked against `ram`
-and `ram+swap` limits. CPU usage for every cgroup is checked against `cpuset.cpus` and `cpu.cfs_period_us` + `cpu.cfs_quota_us` pair assigned for the cgroup. Configuration for the alerts is available in `health.d/cgroups.conf`
-file.
-
-## Monitoring systemd services
-
-Netdata monitors **systemd services**. Example:
-
-![image](https://cloud.githubusercontent.com/assets/2662304/21964372/20cd7b84-db53-11e6-98a2-b9c986b082c0.png)
-
-Support per distribution:
-
-| system | charts shown | `/sys/fs/cgroup` tree | comments |
-|:----------------:|:------------:|:------------------------------------:|:--------------------------|
-| Arch Linux | YES | | |
-| Gentoo | NO | | can be enabled, see below |
-| Ubuntu 16.04 LTS | YES | | |
-| Ubuntu 16.10 | YES | [here](http://pastebin.com/PiWbQEXy) | |
-| Fedora 25 | YES | [here](http://pastebin.com/ax0373wF) | |
-| Debian 8 | NO | | can be enabled, see below |
-| AMI | NO | [here](http://pastebin.com/FrxmptjL) | not a systemd system |
-| CentOS 7.3.1611 | NO | [here](http://pastebin.com/SpzgezAg) | can be enabled, see below |
-
-### Monitored systemd service metrics
-
-- CPU utilization
-- Used memory
-- RSS memory
-- Mapped memory
-- Cache memory
-- Writeback memory
-- Memory minor page faults
-- Memory major page faults
-- Memory charging activity
-- Memory uncharging activity
-- Memory limit failures
-- Swap memory used
-- Disk read bandwidth
-- Disk write bandwidth
-- Disk read operations
-- Disk write operations
-- Throttle disk read bandwidth
-- Throttle disk write bandwidth
-- Throttle disk read operations
-- Throttle disk write operations
-- Queued disk read operations
-- Queued disk write operations
-- Merged disk read operations
-- Merged disk write operations
-
-### How to enable cgroup accounting on systemd systems that is by default disabled
-
-You can verify there is no accounting enabled, by running `systemd-cgtop`. The program will show only resources for
-cgroup `/`, but all services will show nothing.
-
-To enable cgroup accounting, execute this:
-
-```sh
-sed -e 's|^#Default\(.*\)Accounting=.*$|Default\1Accounting=yes|g' /etc/systemd/system.conf >/tmp/system.conf
-```
-
-To see the changes it made, run this:
-
-```sh
-# diff /etc/systemd/system.conf /tmp/system.conf
-40,44c40,44
-< #DefaultCPUAccounting=no
-< #DefaultIOAccounting=no
-< #DefaultBlockIOAccounting=no
-< #DefaultMemoryAccounting=no
-< #DefaultTasksAccounting=yes
----
-> DefaultCPUAccounting=yes
-> DefaultIOAccounting=yes
-> DefaultBlockIOAccounting=yes
-> DefaultMemoryAccounting=yes
-> DefaultTasksAccounting=yes
-```
-
-If you are happy with the changes, run:
-
-```sh
-# copy the file to the right location
-sudo cp /tmp/system.conf /etc/systemd/system.conf
-
-# restart systemd to take it into account
-sudo systemctl daemon-reexec
-```
-
-(`systemctl daemon-reload` does not reload the configuration of the server - so you have to
-execute `systemctl daemon-reexec`).
-
-Now, when you run `systemd-cgtop`, services will start reporting usage (if it does not, restart any service to wake it up). Refresh your Netdata dashboard, and you will have the charts too.
-
-In case memory accounting is missing, you will need to enable it at your kernel, by appending the following kernel boot
-options and rebooting:
-
-```sh
-cgroup_enable=memory swapaccount=1
-```
-
-You can add the above, directly at the `linux` line in your `/boot/grub/grub.cfg` or appending them to
-the `GRUB_CMDLINE_LINUX` in `/etc/default/grub` (in which case you will have to run `update-grub` before rebooting). On
-DigitalOcean debian images you may have to set it at `/etc/default/grub.d/50-cloudimg-settings.cfg`.
-
-Which systemd services are monitored by Netdata is determined by the following pattern list:
-
-```text
-[plugin:cgroups]
- cgroups to match as systemd services = !/system.slice/*/*.service /system.slice/*.service
-```
-
-- - -
-
-## Monitoring ephemeral containers
-
-Netdata monitors containers automatically when it is installed at the host, or when it is installed in a container that
-has access to the `/proc` and `/sys` filesystems of the host.
-
-Network interfaces and cgroups (containers) are self-cleaned. When a network interface or container stops, Netdata might log
-a few errors in error.log complaining about files it cannot find, but immediately:
-
-1. It will detect this is a removed container or network interface
-2. It will freeze/pause all alerts for them
-3. It will mark their charts as obsolete
-4. Obsolete charts are not be offered on new dashboard sessions (so hit F5 and the charts are gone)
-5. Existing dashboard sessions will continue to see them, but of course they will not refresh
-6. Obsolete charts will be removed from memory, 1 hour after the last user viewed them (configurable
- with `[global].cleanup obsolete charts after seconds = 3600` (at `netdata.conf`).
-7. When obsolete charts are removed from memory they are also deleted from disk (configurable
- with `[global].delete obsolete charts files = yes`)
-
-### Monitored container metrics
-
-- CPU usage
-- CPU usage within the limits
-- CPU usage per core
-- Memory usage
-- Writeback memory
-- Memory activity
-- Memory page faults
-- Used memory
-- Used RAM within the limits
-- Memory utilization
-- Memory limit failures
-- I/O bandwidth (all disks)
-- Serviced I/O operations (all disks)
-- Throttle I/O bandwidth (all disks)
-- Throttle serviced I/O operations (all disks)
-- Queued I/O operations (all disks)
-- Merged I/O operations (all disks)
-- CPU pressure
-- Memory pressure
-- Memory full pressure
-- I/O pressure
-- I/O full pressure
-
-Network interfaces are monitored by means of
-the [proc plugin](https://github.com/netdata/netdata/blob/master/collectors/proc.plugin/README.md#monitored-network-interface-metrics).
diff --git a/collectors/cgroups.plugin/metadata.yaml b/collectors/cgroups.plugin/metadata.yaml
deleted file mode 100644
index a1abbb5a9..000000000
--- a/collectors/cgroups.plugin/metadata.yaml
+++ /dev/null
@@ -1,1022 +0,0 @@
-plugin_name: cgroups.plugin
-modules:
- - &module
- meta: &meta
- plugin_name: cgroups.plugin
- module_name: /sys/fs/cgroup
- monitored_instance:
- name: Containers
- link: ""
- categories:
- - data-collection.containers-and-vms
- icon_filename: container.svg
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - containers
- most_popular: true
- overview: &overview
- data_collection: &data_collection
- metrics_description: "Monitor Containers for performance, resource usage, and health status."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: cgroup_10min_cpu_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf
- metric: cgroup.cpu_limit
- info: average cgroup CPU utilization over the last 10 minutes
- - name: cgroup_ram_in_use
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf
- metric: cgroup.mem_usage
- info: cgroup memory utilization
- - name: cgroup_1m_received_packets_rate
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf
- metric: cgroup.net_packets
- info: average number of packets received by the network interface ${label:device} over the last minute
- - name: cgroup_10s_received_packets_storm
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf
- metric: cgroup.net_packets
- info:
- ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over
- the last minute
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: cgroup
- description: ""
- labels:
- - name: container_name
- description: The container name or group path if name resolution fails.
- - name: image
- description: Docker/Podman container image name.
- metrics:
- - name: cgroup.cpu_limit
- description: CPU Usage within the limits
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: used
- - name: cgroup.cpu
- description: CPU Usage (100% = 1 core)
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: user
- - name: system
- - name: cgroup.cpu_per_core
- description: CPU Usage (100% = 1 core) Per Core
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: a dimension per core
- - name: cgroup.throttled
- description: CPU Throttled Runnable Periods
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: throttled
- - name: cgroup.throttled_duration
- description: CPU Throttled Time Duration
- unit: "ms"
- chart_type: line
- dimensions:
- - name: duration
- - name: cgroup.cpu_shares
- description: CPU Time Relative Share
- unit: "shares"
- chart_type: line
- dimensions:
- - name: shares
- - name: cgroup.mem
- description: Memory Usage
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: cache
- - name: rss
- - name: swap
- - name: rss_huge
- - name: mapped_file
- - name: cgroup.writeback
- description: Writeback Memory
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: dirty
- - name: writeback
- - name: cgroup.mem_activity
- description: Memory Activity
- unit: "MiB/s"
- chart_type: line
- dimensions:
- - name: in
- - name: out
- - name: cgroup.pgfaults
- description: Memory Page Faults
- unit: "MiB/s"
- chart_type: line
- dimensions:
- - name: pgfault
- - name: swap
- - name: cgroup.mem_usage
- description: Used Memory
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: ram
- - name: swap
- - name: cgroup.mem_usage_limit
- description: Used RAM within the limits
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: available
- - name: used
- - name: cgroup.mem_utilization
- description: Memory Utilization
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: utilization
- - name: cgroup.mem_failcnt
- description: Memory Limit Failures
- unit: "count"
- chart_type: line
- dimensions:
- - name: failures
- - name: cgroup.io
- description: I/O Bandwidth (all disks)
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: cgroup.serviced_ops
- description: Serviced I/O Operations (all disks)
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: cgroup.throttle_io
- description: Throttle I/O Bandwidth (all disks)
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: cgroup.throttle_serviced_ops
- description: Throttle Serviced I/O Operations (all disks)
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: cgroup.queued_ops
- description: Queued I/O Operations (all disks)
- unit: "operations"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: cgroup.merged_ops
- description: Merged I/O Operations (all disks)
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: cgroup.cpu_some_pressure
- description: CPU some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: cgroup.cpu_some_pressure_stall_time
- description: CPU some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: cgroup.cpu_full_pressure
- description: CPU full pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: cgroup.cpu_full_pressure_stall_time
- description: CPU full pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: cgroup.memory_some_pressure
- description: Memory some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: cgroup.memory_some_pressure_stall_time
- description: Memory some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: cgroup.memory_full_pressure
- description: Memory full pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: cgroup.memory_full_pressure_stall_time
- description: Memory full pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: cgroup.io_some_pressure
- description: I/O some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: cgroup.io_some_pressure_stall_time
- description: I/O some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: cgroup.io_full_pressure
- description: I/O some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: cgroup.io_full_pressure_stall_time
- description: I/O some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: cgroup.pids_current
- description: Number of processes
- unit: "pids"
- chart_type: line
- dimensions:
- - name: pids
- - name: cgroup network device
- description: ""
- labels:
- - name: container_name
- description: The container name or group path if name resolution fails.
- - name: image
- description: Docker/Podman container image name.
- - name: device
- description: "The name of the host network interface linked to the container's network interface."
- - name: container_device
- description: Container network interface name.
- - name: interface_type
- description: 'Network interface type. Always "virtual" for the containers.'
- metrics:
- - name: cgroup.net_net
- description: Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: cgroup.net_packets
- description: Packets
- unit: "pps"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: multicast
- - name: cgroup.net_errors
- description: Interface Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: inbound
- - name: outbound
- - name: cgroup.net_drops
- description: Interface Drops
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: inbound
- - name: outbound
- - name: cgroup.net_fifo
- description: Interface FIFO Buffer Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: receive
- - name: transmit
- - name: cgroup.net_compressed
- description: Interface FIFO Buffer Errors
- unit: "pps"
- chart_type: line
- dimensions:
- - name: receive
- - name: sent
- - name: cgroup.net_events
- description: Network Interface Events
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: frames
- - name: collisions
- - name: carrier
- - name: cgroup.net_operstate
- description: Interface Operational State
- unit: "state"
- chart_type: line
- dimensions:
- - name: up
- - name: down
- - name: notpresent
- - name: lowerlayerdown
- - name: testing
- - name: dormant
- - name: unknown
- - name: cgroup.net_carrier
- description: Interface Physical Link State
- unit: "state"
- chart_type: line
- dimensions:
- - name: up
- - name: down
- - name: cgroup.net_mtu
- description: Interface MTU
- unit: "octets"
- chart_type: line
- dimensions:
- - name: mtu
- - <<: *module
- meta:
- <<: *meta
- monitored_instance:
- name: Kubernetes Containers
- link: https://kubernetes.io/
- icon_filename: kubernetes.svg
- categories:
- #- data-collection.containers-and-vms
- - data-collection.kubernetes
- keywords:
- - k8s
- - kubernetes
- - pods
- - containers
- overview:
- <<: *overview
- data-collection:
- <<: *data_collection
- metrics_description: Monitor Kubernetes Clusters for performance, resource usage, and health status.
- alerts:
- - name: k8s_cgroup_10min_cpu_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf
- metric: k8s.cgroup.cpu_limit
- info: average cgroup CPU utilization over the last 10 minutes
- - name: k8s_cgroup_ram_in_use
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf
- metric: k8s.cgroup.mem_usage
- info: cgroup memory utilization
- - name: k8s_cgroup_1m_received_packets_rate
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf
- metric: k8s.cgroup.net_packets
- info: average number of packets received by the network interface ${label:device} over the last minute
- - name: k8s_cgroup_10s_received_packets_storm
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf
- metric: k8s.cgroup.net_packets
- info:
- ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over
- the last minute
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: k8s cgroup
- description: These metrics refer to the Pod container.
- labels:
- - name: k8s_node_name
- description: 'Node name. The value of _pod.spec.nodeName_.'
- - name: k8s_namespace
- description: 'Namespace name. The value of _pod.metadata.namespace_.'
- - name: k8s_controller_kind
- description: 'Controller kind (ReplicaSet, DaemonSet, StatefulSet, Job, etc.). The value of _pod.OwnerReferences.Controller.Kind_.'
- - name: k8s_controller_name
- description: 'Controller name.The value of _pod.OwnerReferences.Controller.Name_.'
- - name: k8s_pod_name
- description: 'Pod name. The value of _pod.metadata.name_.'
- - name: k8s_container_name
- description: 'Container name. The value of _pod.spec.containers.name_.'
- - name: k8s_kind
- description: 'Instance kind: "pod" or "container".'
- - name: k8s_qos_class
- description: 'QoS class (guaranteed, burstable, besteffort).'
- - name: k8s_cluster_id
- description: 'Cluster ID. The value of kube-system namespace _namespace.metadata.uid_.'
- metrics:
- - name: k8s.cgroup.cpu_limit
- description: CPU Usage within the limits
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: used
- - name: k8s.cgroup.cpu
- description: CPU Usage (100% = 1000 mCPU)
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: user
- - name: system
- - name: k8s.cgroup.cpu_per_core
- description: CPU Usage (100% = 1000 mCPU) Per Core
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: a dimension per core
- - name: k8s.cgroup.throttled
- description: CPU Throttled Runnable Periods
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: throttled
- - name: k8s.cgroup.throttled_duration
- description: CPU Throttled Time Duration
- unit: "ms"
- chart_type: line
- dimensions:
- - name: duration
- - name: k8s.cgroup.cpu_shares
- description: CPU Time Relative Share
- unit: "shares"
- chart_type: line
- dimensions:
- - name: shares
- - name: k8s.cgroup.mem
- description: Memory Usage
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: cache
- - name: rss
- - name: swap
- - name: rss_huge
- - name: mapped_file
- - name: k8s.cgroup.writeback
- description: Writeback Memory
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: dirty
- - name: writeback
- - name: k8s.cgroup.mem_activity
- description: Memory Activity
- unit: "MiB/s"
- chart_type: line
- dimensions:
- - name: in
- - name: out
- - name: k8s.cgroup.pgfaults
- description: Memory Page Faults
- unit: "MiB/s"
- chart_type: line
- dimensions:
- - name: pgfault
- - name: swap
- - name: k8s.cgroup.mem_usage
- description: Used Memory
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: ram
- - name: swap
- - name: k8s.cgroup.mem_usage_limit
- description: Used RAM within the limits
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: available
- - name: used
- - name: k8s.cgroup.mem_utilization
- description: Memory Utilization
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: utilization
- - name: k8s.cgroup.mem_failcnt
- description: Memory Limit Failures
- unit: "count"
- chart_type: line
- dimensions:
- - name: failures
- - name: k8s.cgroup.io
- description: I/O Bandwidth (all disks)
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: k8s.cgroup.serviced_ops
- description: Serviced I/O Operations (all disks)
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: k8s.cgroup.throttle_io
- description: Throttle I/O Bandwidth (all disks)
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: k8s.cgroup.throttle_serviced_ops
- description: Throttle Serviced I/O Operations (all disks)
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: k8s.cgroup.queued_ops
- description: Queued I/O Operations (all disks)
- unit: "operations"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: k8s.cgroup.merged_ops
- description: Merged I/O Operations (all disks)
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: k8s.cgroup.cpu_some_pressure
- description: CPU some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: k8s.cgroup.cpu_some_pressure_stall_time
- description: CPU some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: k8s.cgroup.cpu_full_pressure
- description: CPU full pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: k8s.cgroup.cpu_full_pressure_stall_time
- description: CPU full pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: k8s.cgroup.memory_some_pressure
- description: Memory some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: k8s.cgroup.memory_some_pressure_stall_time
- description: Memory some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: k8s.cgroup.memory_full_pressure
- description: Memory full pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: k8s.cgroup.memory_full_pressure_stall_time
- description: Memory full pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: k8s.cgroup.io_some_pressure
- description: I/O some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: k8s.cgroup.io_some_pressure_stall_time
- description: I/O some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: k8s.cgroup.io_full_pressure
- description: I/O some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: k8s.cgroup.io_full_pressure_stall_time
- description: I/O some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: k8s.cgroup.pids_current
- description: Number of processes
- unit: "pids"
- chart_type: line
- dimensions:
- - name: pids
- - name: k8s cgroup network device
- description: These metrics refer to the Pod container network interface.
- labels:
- - name: device
- description: "The name of the host network interface linked to the container's network interface."
- - name: container_device
- description: Container network interface name.
- - name: interface_type
- description: 'Network interface type. Always "virtual" for the containers.'
- - name: k8s_node_name
- description: 'Node name. The value of _pod.spec.nodeName_.'
- - name: k8s_namespace
- description: 'Namespace name. The value of _pod.metadata.namespace_.'
- - name: k8s_controller_kind
- description: 'Controller kind (ReplicaSet, DaemonSet, StatefulSet, Job, etc.). The value of _pod.OwnerReferences.Controller.Kind_.'
- - name: k8s_controller_name
- description: 'Controller name.The value of _pod.OwnerReferences.Controller.Name_.'
- - name: k8s_pod_name
- description: 'Pod name. The value of _pod.metadata.name_.'
- - name: k8s_container_name
- description: 'Container name. The value of _pod.spec.containers.name_.'
- - name: k8s_kind
- description: 'Instance kind: "pod" or "container".'
- - name: k8s_qos_class
- description: 'QoS class (guaranteed, burstable, besteffort).'
- - name: k8s_cluster_id
- description: 'Cluster ID. The value of kube-system namespace _namespace.metadata.uid_.'
- metrics:
- - name: k8s.cgroup.net_net
- description: Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: k8s.cgroup.net_packets
- description: Packets
- unit: "pps"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: multicast
- - name: k8s.cgroup.net_errors
- description: Interface Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: inbound
- - name: outbound
- - name: k8s.cgroup.net_drops
- description: Interface Drops
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: inbound
- - name: outbound
- - name: k8s.cgroup.net_fifo
- description: Interface FIFO Buffer Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: receive
- - name: transmit
- - name: k8s.cgroup.net_compressed
- description: Interface FIFO Buffer Errors
- unit: "pps"
- chart_type: line
- dimensions:
- - name: receive
- - name: sent
- - name: k8s.cgroup.net_events
- description: Network Interface Events
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: frames
- - name: collisions
- - name: carrier
- - name: k8s.cgroup.net_operstate
- description: Interface Operational State
- unit: "state"
- chart_type: line
- dimensions:
- - name: up
- - name: down
- - name: notpresent
- - name: lowerlayerdown
- - name: testing
- - name: dormant
- - name: unknown
- - name: k8s.cgroup.net_carrier
- description: Interface Physical Link State
- unit: "state"
- chart_type: line
- dimensions:
- - name: up
- - name: down
- - name: k8s.cgroup.net_mtu
- description: Interface MTU
- unit: "octets"
- chart_type: line
- dimensions:
- - name: mtu
- - <<: *module
- meta:
- <<: *meta
- monitored_instance:
- name: Systemd Services
- link: ""
- icon_filename: systemd.svg
- categories:
- - data-collection.systemd
- keywords:
- - systemd
- - services
- overview:
- <<: *overview
- data-collection:
- <<: *data_collection
- metrics_desctiption: "Monitor Systemd Services for performance, resource usage, and health status."
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: systemd service
- description: ""
- labels:
- - name: service_name
- description: Service name
- metrics:
- - name: systemd.service.cpu.utilization
- description: Systemd Services CPU utilization (100% = 1 core)
- unit: percentage
- chart_type: stacked
- dimensions:
- - name: user
- - name: system
- - name: systemd.service.memory.usage
- description: Systemd Services Used Memory
- unit: MiB
- chart_type: stacked
- dimensions:
- - name: ram
- - name: swap
- - name: systemd.service.memory.failcnt
- description: Systemd Services Memory Limit Failures
- unit: failures/s
- chart_type: line
- dimensions:
- - name: fail
- - name: systemd.service.memory.ram.usage
- description: Systemd Services Memory
- unit: MiB
- chart_type: stacked
- dimensions:
- - name: rss
- - name: cache
- - name: mapped_file
- - name: rss_huge
- - name: systemd.service.memory.writeback
- description: Systemd Services Writeback Memory
- unit: MiB
- chart_type: stacked
- dimensions:
- - name: writeback
- - name: dirty
- - name: systemd.service.memory.paging.faults
- description: Systemd Services Memory Minor and Major Page Faults
- unit: MiB/s
- chart_type: area
- dimensions:
- - name: minor
- - name: major
- - name: systemd.service.memory.paging.io
- description: Systemd Services Memory Paging IO
- unit: MiB/s
- chart_type: area
- dimensions:
- - name: in
- - name: out
- - name: systemd.service.disk.io
- description: Systemd Services Disk Read/Write Bandwidth
- unit: KiB/s
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: systemd.service.disk.iops
- description: Systemd Services Disk Read/Write Operations
- unit: operations/s
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: systemd.service.disk.throttle.io
- description: Systemd Services Throttle Disk Read/Write Bandwidth
- unit: KiB/s
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: systemd.service.disk.throttle.iops
- description: Systemd Services Throttle Disk Read/Write Operations
- unit: operations/s
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: systemd.service.disk.queued_iops
- description: Systemd Services Queued Disk Read/Write Operations
- unit: operations/s
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: systemd.service.disk.merged_iops
- description: Systemd Services Merged Disk Read/Write Operations
- unit: operations/s
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: systemd.service.pids.current
- description: Systemd Services Number of Processes
- unit: pids
- chart_type: line
- dimensions:
- - name: pids
- - <<: *module
- meta:
- <<: *meta
- monitored_instance:
- name: Virtual Machines
- link: ""
- icon_filename: container.svg
- categories:
- - data-collection.containers-and-vms
- keywords:
- - vms
- - virtualization
- - container
- overview:
- <<: *overview
- data_collection:
- <<: *data_collection
- metrics_description: "Monitor Virtual Machines for performance, resource usage, and health status."
- - <<: *module
- meta:
- <<: *meta
- monitored_instance:
- name: LXC Containers
- link: ""
- icon_filename: lxc.png
- categories:
- - data-collection.containers-and-vms
- keywords:
- - lxc
- - lxd
- - container
- overview:
- <<: *overview
- data_collection:
- <<: *data_collection
- metrics_description: "Monitor LXC Containers for performance, resource usage, and health status."
- - <<: *module
- meta:
- <<: *meta
- monitored_instance:
- name: Libvirt Containers
- link: ""
- icon_filename: libvirt.png
- categories:
- - data-collection.containers-and-vms
- keywords:
- - libvirt
- - container
- overview:
- <<: *overview
- data_collection:
- <<: *data_collection
- metrics_description: "Monitor Libvirt for performance, resource usage, and health status."
- - <<: *module
- meta:
- <<: *meta
- monitored_instance:
- name: oVirt Containers
- link: ""
- icon_filename: ovirt.svg
- categories:
- - data-collection.containers-and-vms
- keywords:
- - ovirt
- - container
- overview:
- <<: *overview
- data_collection:
- <<: *data_collection
- metrics_description: "Monitor oVirt for performance, resource usage, and health status."
- - <<: *module
- meta:
- <<: *meta
- monitored_instance:
- name: Proxmox Containers
- link: ""
- icon_filename: proxmox.png
- categories:
- - data-collection.containers-and-vms
- keywords:
- - proxmox
- - container
- overview:
- <<: *overview
- data_collection:
- <<: *data_collection
- metrics_description: "Monitor Proxmox for performance, resource usage, and health status."
diff --git a/collectors/charts.d.plugin/Makefile.am b/collectors/charts.d.plugin/Makefile.am
deleted file mode 100644
index f82992fd4..000000000
--- a/collectors/charts.d.plugin/Makefile.am
+++ /dev/null
@@ -1,49 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-CLEANFILES = \
- charts.d.plugin \
- $(NULL)
-
-include $(top_srcdir)/build/subst.inc
-SUFFIXES = .in
-
-dist_libconfig_DATA = \
- charts.d.conf \
- $(NULL)
-
-dist_plugins_SCRIPTS = \
- charts.d.dryrun-helper.sh \
- charts.d.plugin \
- loopsleepms.sh.inc \
- $(NULL)
-
-dist_noinst_DATA = \
- charts.d.plugin.in \
- README.md \
- $(NULL)
-
-dist_charts_SCRIPTS = \
- $(NULL)
-
-dist_charts_DATA = \
- $(NULL)
-
-userchartsconfigdir=$(configdir)/charts.d
-dist_userchartsconfig_DATA = \
- $(NULL)
-
-# Explicitly install directories to avoid permission issues due to umask
-install-exec-local:
- $(INSTALL) -d $(DESTDIR)$(userchartsconfigdir)
-
-chartsconfigdir=$(libconfigdir)/charts.d
-dist_chartsconfig_DATA = \
- $(NULL)
-
-include ap/Makefile.inc
-include apcupsd/Makefile.inc
-include example/Makefile.inc
-include libreswan/Makefile.inc
-include opensips/Makefile.inc
-include sensors/Makefile.inc
diff --git a/collectors/charts.d.plugin/README.md b/collectors/charts.d.plugin/README.md
deleted file mode 100644
index 97c2446fa..000000000
--- a/collectors/charts.d.plugin/README.md
+++ /dev/null
@@ -1,190 +0,0 @@
-# charts.d.plugin
-
-`charts.d.plugin` is a Netdata external plugin. It is an **orchestrator** for data collection modules written in `BASH` v4+.
-
-1. It runs as an independent process `ps fax` shows it
-2. It is started and stopped automatically by Netdata
-3. It communicates with Netdata via a unidirectional pipe (sending data to the `netdata` daemon)
-4. Supports any number of data collection **modules**
-
-To better understand the guidelines and the API behind our External plugins, please have a look at the [Introduction to External plugins](https://github.com/netdata/netdata/blob/master/collectors/plugins.d/README.md) prior to reading this page.
-
-
-`charts.d.plugin` has been designed so that the actual script that will do data collection will be permanently in
-memory, collecting data with as little overheads as possible
-(i.e. initialize once, repeatedly collect values with minimal overhead).
-
-`charts.d.plugin` looks for scripts in `/usr/lib/netdata/charts.d`.
-The scripts should have the filename suffix: `.chart.sh`.
-
-By default, `charts.d.plugin` is not included as part of the install when using [our official native DEB/RPM packages](https://github.com/netdata/netdata/blob/master/packaging/installer/methods/packages.md). You can install it by installing the `netdata-plugin-chartsd` package.
-
-## Configuration
-
-`charts.d.plugin` itself can be [configured](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#use-edit-config-to-edit-configuration-files) using the configuration file `/etc/netdata/charts.d.conf`. This file is also a BASH script.
-
-In this file, you can place statements like this:
-
-```conf
-enable_all_charts="yes"
-X="yes"
-Y="no"
-```
-
-where `X` and `Y` are the names of individual charts.d collector scripts.
-When set to `yes`, charts.d will evaluate the collector script (see below).
-When set to `no`, charts.d will ignore the collector script.
-
-The variable `enable_all_charts` sets the default enable/disable state for all charts.
-
-## A charts.d module
-
-A `charts.d.plugin` module is a BASH script defining a few functions.
-
-For a module called `X`, the following criteria must be met:
-
-1. The module script must be called `X.chart.sh` and placed in `/usr/libexec/netdata/charts.d`.
-
-2. If the module needs a configuration, it should be called `X.conf` and placed in `/etc/netdata/charts.d`.
- The configuration file `X.conf` is also a BASH script itself.
- You can edit the default files supplied by Netdata, by editing `/etc/netdata/edit-config charts.d/X.conf`, where `X` is the name of the module.
-
-3. All functions and global variables defined in the script and its configuration, must begin with `X_`.
-
-4. The following functions must be defined:
-
- - `X_check()` - returns 0 or 1 depending on whether the module is able to run or not
- (following the standard Linux command line return codes: 0 = OK, the collector can operate and 1 = FAILED,
- the collector cannot be used).
-
- - `X_create()` - creates the Netdata charts (commands `CHART` and `DIMENSION`).
- The return value does matter: 0 = OK, 1 = FAILED.
-
- - `X_update()` - collects the values for the defined charts (commands `BEGIN`, `SET`, `END`).
- The return value also matters: 0 = OK, 1 = FAILED.
-
-5. The following global variables are available to be set:
- - `X_update_every` - is the data collection frequency for the module script, in seconds.
-
-The module script may use more functions or variables. But all of them must begin with `X_`.
-
-### X_check()
-
-The purpose of the BASH function `X_check()` is to check if the module can collect data (or check its config).
-
-For example, if the module is about monitoring a local mysql database, the `X_check()` function may attempt to
-connect to a local mysql database to find out if it can read the values it needs.
-
-`X_check()` is run only once for the lifetime of the module.
-
-### X_create()
-
-The purpose of the BASH function `X_create()` is to create the charts and dimensions using the standard Netdata
-plugin guidelines.
-
-`X_create()` will be called just once and only after `X_check()` was successful.
-You can however call it yourself when there is need for it (for example to add a new dimension to an existing chart).
-
-A non-zero return value will disable the collector.
-
-### X_update()
-
-`X_update()` will be called repeatedly every `X_update_every` seconds, to collect new values and send them to Netdata,
-following the Netdata plugin guidelines.
-
-The function will be called with one parameter: microseconds since the last time it was run. This value should be
-appended to the `BEGIN` statement of every chart updated by the collector script.
-
-A non-zero return value will disable the collector.
-
-### Useful functions charts.d provides
-
-Module scripts can use the following charts.d functions:
-
-#### require_cmd command
-
-`require_cmd()` will check if a command is available in the running system.
-
-For example, your `X_check()` function may use it like this:
-
-```sh
-mysql_check() {
- require_cmd mysql || return 1
- return 0
-}
-```
-
-Using the above, if the command `mysql` is not available in the system, the `mysql` module will be disabled.
-
-#### fixid "string"
-
-`fixid()` will get a string and return a properly formatted id for a chart or dimension.
-
-This is an expensive function that should not be used in `X_update()`.
-You can keep the generated id in a BASH associative array to have the values availables in `X_update()`, like this:
-
-```sh
-declare -A X_ids=()
-X_create() {
- local name="a very bad name for id"
-
- X_ids[$name]="$(fixid "$name")"
-}
-
-X_update() {
- local microseconds="$1"
-
- ...
- local name="a very bad name for id"
- ...
-
- echo "BEGIN ${X_ids[$name]} $microseconds"
- ...
-}
-```
-
-### Debugging your collectors
-
-You can run `charts.d.plugin` by hand with something like this:
-
-```sh
-# become user netdata
-sudo su -s /bin/sh netdata
-
-# run the plugin in debug mode
-/usr/libexec/netdata/plugins.d/charts.d.plugin debug 1 X Y Z
-```
-
-Charts.d will run in `debug` mode, with an update frequency of `1`, evaluating only the collector scripts
-`X`, `Y` and `Z`. You can define zero or more module scripts. If none is defined, charts.d will evaluate all
-module scripts available.
-
-Keep in mind that if your configs are not in `/etc/netdata`, you should do the following before running
-`charts.d.plugin`:
-
-```sh
-export NETDATA_USER_CONFIG_DIR="/path/to/etc/netdata"
-```
-
-Also, remember that Netdata runs `chart.d.plugin` as user `netdata` (or any other user the `netdata` process is configured to run as).
-
-## Running multiple instances of charts.d.plugin
-
-`charts.d.plugin` will call the `X_update()` function one after another. This means that a delay in collector `X`
-will also delay the collection of `Y` and `Z`.
-
-You can have multiple `charts.d.plugin` running to overcome this problem.
-
-This is what you need to do:
-
-1. Decide a new name for the new charts.d instance: example `charts2.d`.
-
-2. Create/edit the files `/etc/netdata/charts.d.conf` and `/etc/netdata/charts2.d.conf` and enable / disable the
- module you want each to run. Remember to set `enable_all_charts="no"` to both of them, and enable the individual
- modules for each.
-
-3. link `/usr/libexec/netdata/plugins.d/charts.d.plugin` to `/usr/libexec/netdata/plugins.d/charts2.d.plugin`.
- Netdata will spawn a new charts.d process.
-
-Execute the above in this order, since Netdata will (by default) attempt to start new plugins soon after they are
-created in `/usr/libexec/netdata/plugins.d/`.
diff --git a/collectors/charts.d.plugin/ap/Makefile.inc b/collectors/charts.d.plugin/ap/Makefile.inc
deleted file mode 100644
index a2dd375ac..000000000
--- a/collectors/charts.d.plugin/ap/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_charts_DATA += ap/ap.chart.sh
-dist_chartsconfig_DATA += ap/ap.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += ap/README.md ap/Makefile.inc
-
diff --git a/collectors/charts.d.plugin/apcupsd/Makefile.inc b/collectors/charts.d.plugin/apcupsd/Makefile.inc
deleted file mode 100644
index 19cb9cad7..000000000
--- a/collectors/charts.d.plugin/apcupsd/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_charts_DATA += apcupsd/apcupsd.chart.sh
-dist_chartsconfig_DATA += apcupsd/apcupsd.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += apcupsd/README.md apcupsd/Makefile.inc
-
diff --git a/collectors/charts.d.plugin/apcupsd/metadata.yaml b/collectors/charts.d.plugin/apcupsd/metadata.yaml
deleted file mode 100644
index c333dc964..000000000
--- a/collectors/charts.d.plugin/apcupsd/metadata.yaml
+++ /dev/null
@@ -1,256 +0,0 @@
-plugin_name: charts.d.plugin
-modules:
- - meta:
- plugin_name: charts.d.plugin
- module_name: apcupsd
- monitored_instance:
- name: APC UPS
- link: "https://www.apc.com"
- categories:
- - data-collection.ups
- icon_filename: "apc.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - ups
- - apc
- - power
- - supply
- - battery
- - apcupsd
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor APC UPS performance with Netdata for optimal uninterruptible power supply operations. Enhance your power supply reliability with real-time APC UPS metrics."
- method_description: "The collector uses the `apcaccess` tool to contact the `apcupsd` daemon and get the APC UPS statistics."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: "By default, with no configuration provided, the collector will try to contact 127.0.0.1:3551 with using the `apcaccess` utility."
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: "Install charts.d plugin"
- description: |
- If [using our official native DEB/RPM packages](https://github.com/netdata/netdata/blob/master/packaging/installer/UPDATE.md#determine-which-installation-method-you-used), make sure `netdata-plugin-chartsd` is installed.
- - title: "Required software"
- description: "Make sure the `apcaccess` and `apcupsd` are installed and running."
- configuration:
- file:
- name: charts.d/apcupsd.conf
- options:
- description: |
- The config file is sourced by the charts.d plugin. It's a standard bash file.
-
- The following collapsed table contains all the options that can be configured for the apcupsd collector.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: apcupsd_sources
- description: This is an array of apcupsd sources. You can have multiple entries there. Please refer to the example below on how to set it.
- default_value: "127.0.0.1:3551"
- required: false
- - name: apcupsd_timeout
- description: How long to wait for apcupsd to respond.
- default_value: 3
- required: false
- - name: apcupsd_update_every
- description: The data collection frequency. If unset, will inherit the netdata update frequency.
- default_value: 1
- required: false
- - name: apcupsd_priority
- description: The charts priority on the dashboard.
- default_value: 90000
- required: false
- - name: apcupsd_retries
- description: The number of retries to do in case of failure before disabling the collector.
- default_value: 10
- required: false
- examples:
- folding:
- enabled: false
- title: "Config"
- list:
- - name: Multiple apcupsd sources
- description: Specify a multiple apcupsd sources along with a custom update interval
- config: |
- # add all your APC UPSes in this array - uncomment it too
- declare -A apcupsd_sources=(
- ["local"]="127.0.0.1:3551",
- ["remote"]="1.2.3.4:3551"
- )
-
- # how long to wait for apcupsd to respond
- #apcupsd_timeout=3
-
- # the data collection frequency
- # if unset, will inherit the netdata update frequency
- apcupsd_update_every=5
-
- # the charts priority on the dashboard
- #apcupsd_priority=90000
-
- # the number of retries to do in case of failure
- # before disabling the module
- #apcupsd_retries=10
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: apcupsd_ups_charge
- link: https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf
- metric: apcupsd.charge
- info: average UPS charge over the last minute
- os: "*"
- - name: apcupsd_10min_ups_load
- link: https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf
- metric: apcupsd.load
- info: average UPS load over the last 10 minutes
- os: "*"
- - name: apcupsd_last_collected_secs
- link: https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf
- metric: apcupsd.load
- info: number of seconds since the last successful data collection
- - name: apcupsd_selftest_warning
- link: https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf
- metric: apcupsd.selftest
- info: self-test failed due to insufficient battery capacity or due to overload.
- - name: apcupsd_status_onbatt
- link: https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf
- metric: apcupsd.status
- info: APC UPS has switched to battery power because the input power has failed
- - name: apcupsd_status_overload
- link: https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf
- metric: apcupsd.status
- info: APC UPS is overloaded and cannot supply enough power to the load
- - name: apcupsd_status_lowbatt
- link: https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf
- metric: apcupsd.status
- info: APC UPS battery is low and needs to be recharged
- - name: apcupsd_status_replacebatt
- link: https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf
- metric: apcupsd.status
- info: APC UPS battery has reached the end of its lifespan and needs to be replaced
- - name: apcupsd_status_nobatt
- link: https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf
- metric: apcupsd.status
- info: APC UPS has no battery
- - name: apcupsd_status_commlost
- link: https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf
- metric: apcupsd.status
- info: APC UPS communication link is lost
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: ups
- description: "Metrics related to UPS. Each UPS provides its own set of the following metrics."
- labels: []
- metrics:
- - name: apcupsd.charge
- description: UPS Charge
- unit: "percentage"
- chart_type: area
- dimensions:
- - name: charge
- - name: apcupsd.battery.voltage
- description: UPS Battery Voltage
- unit: "Volts"
- chart_type: line
- dimensions:
- - name: voltage
- - name: nominal
- - name: apcupsd.input.voltage
- description: UPS Input Voltage
- unit: "Volts"
- chart_type: line
- dimensions:
- - name: voltage
- - name: min
- - name: max
- - name: apcupsd.output.voltage
- description: UPS Output Voltage
- unit: "Volts"
- chart_type: line
- dimensions:
- - name: absolute
- - name: nominal
- - name: apcupsd.input.frequency
- description: UPS Input Voltage
- unit: "Hz"
- chart_type: line
- dimensions:
- - name: frequency
- - name: apcupsd.load
- description: UPS Load
- unit: "percentage"
- chart_type: area
- dimensions:
- - name: load
- - name: apcupsd.load_usage
- description: UPS Load Usage
- unit: "Watts"
- chart_type: area
- dimensions:
- - name: load
- - name: apcupsd.temperature
- description: UPS Temperature
- unit: "Celsius"
- chart_type: line
- dimensions:
- - name: temp
- - name: apcupsd.time
- description: UPS Time Remaining
- unit: "Minutes"
- chart_type: area
- dimensions:
- - name: time
- - name: apcupsd.online
- description: UPS ONLINE flag
- unit: "boolean"
- chart_type: line
- dimensions:
- - name: online
- - name: apcupsd.selftest
- description: UPS Self-Test status
- unit: status
- chart_type: line
- dimensions:
- - name: OK
- - name: NO
- - name: BT
- - name: NG
- - name: apcupsd.status
- description: UPS Status
- unit: status
- chart_type: line
- dimensions:
- - name: ONLINE
- - name: ONBATT
- - name: OVERLOAD
- - name: LOWBATT
- - name: REPLACEBATT
- - name: NOBATT
- - name: SLAVE
- - name: SLAVEDOWN
- - name: COMMLOST
- - name: CAL
- - name: TRIM
- - name: BOOST
- - name: SHUTTING_DOWN
diff --git a/collectors/charts.d.plugin/example/Makefile.inc b/collectors/charts.d.plugin/example/Makefile.inc
deleted file mode 100644
index e6838fbbe..000000000
--- a/collectors/charts.d.plugin/example/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_charts_DATA += example/example.chart.sh
-dist_chartsconfig_DATA += example/example.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += example/README.md example/Makefile.inc
-
diff --git a/collectors/charts.d.plugin/example/README.md b/collectors/charts.d.plugin/example/README.md
deleted file mode 100644
index c2860eb3d..000000000
--- a/collectors/charts.d.plugin/example/README.md
+++ /dev/null
@@ -1,14 +0,0 @@
-<!--
-title: "Example"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/example/README.md"
-sidebar_label: "example-charts.d.plugin"
-learn_status: "Published"
-learn_topic_type: "References"
-learn_rel_path: "Integrations/Monitor/Mock Collectors"
--->
-
-# Example
-
-If you want to understand how charts.d data collector functions, check out the [charts.d example](https://raw.githubusercontent.com/netdata/netdata/master/collectors/charts.d.plugin/example/example.chart.sh).
-
-
diff --git a/collectors/charts.d.plugin/libreswan/Makefile.inc b/collectors/charts.d.plugin/libreswan/Makefile.inc
deleted file mode 100644
index af767d0dd..000000000
--- a/collectors/charts.d.plugin/libreswan/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_charts_DATA += libreswan/libreswan.chart.sh
-dist_chartsconfig_DATA += libreswan/libreswan.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += libreswan/README.md libreswan/Makefile.inc
-
diff --git a/collectors/charts.d.plugin/opensips/Makefile.inc b/collectors/charts.d.plugin/opensips/Makefile.inc
deleted file mode 100644
index a7b5d3a92..000000000
--- a/collectors/charts.d.plugin/opensips/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_charts_DATA += opensips/opensips.chart.sh
-dist_chartsconfig_DATA += opensips/opensips.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += opensips/README.md opensips/Makefile.inc
-
diff --git a/collectors/charts.d.plugin/sensors/Makefile.inc b/collectors/charts.d.plugin/sensors/Makefile.inc
deleted file mode 100644
index f466a1b62..000000000
--- a/collectors/charts.d.plugin/sensors/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_charts_DATA += sensors/sensors.chart.sh
-dist_chartsconfig_DATA += sensors/sensors.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += sensors/README.md sensors/Makefile.inc
-
diff --git a/collectors/charts.d.plugin/sensors/metadata.yaml b/collectors/charts.d.plugin/sensors/metadata.yaml
deleted file mode 100644
index 47f6f4042..000000000
--- a/collectors/charts.d.plugin/sensors/metadata.yaml
+++ /dev/null
@@ -1,182 +0,0 @@
-plugin_name: charts.d.plugin
-modules:
- - meta:
- plugin_name: charts.d.plugin
- module_name: sensors
- monitored_instance:
- name: Linux Sensors (sysfs)
- link: "https://www.kernel.org/doc/Documentation/hwmon/sysfs-interface"
- categories:
- - data-collection.hardware-devices-and-sensors
- icon_filename: "microchip.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - sensors
- - sysfs
- - hwmon
- - rpi
- - raspberry pi
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- Use this collector when `lm-sensors` doesn't work on your device (e.g. for RPi temperatures).
- For all other cases use the [Python collector](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/sensors), which supports multiple jobs, is more efficient and performs calculations on top of the kernel provided values."
- method_description: |
- It will provide charts for all configured system sensors, by reading sensors directly from the kernel.
- The values graphed are the raw hardware values of the sensors.
- supported_platforms:
- include: [Linux]
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: "By default, the collector will try to read entries under `/sys/devices`"
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: "Install charts.d plugin"
- description: |
- If [using our official native DEB/RPM packages](https://github.com/netdata/netdata/blob/master/packaging/installer/UPDATE.md#determine-which-installation-method-you-used), make sure `netdata-plugin-chartsd` is installed.
- - title: "Enable the sensors collector"
- description: |
- The `sensors` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`, to edit the `charts.d.conf` file.
-
- ```bash
- cd /etc/netdata # Replace this path with your Netdata config directory, if different
- sudo ./edit-config charts.d.conf
- ```
-
- Change the value of the `sensors` setting to `force` and uncomment the line. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
- configuration:
- file:
- name: charts.d/sensors.conf
- options:
- description: |
- The config file is sourced by the charts.d plugin. It's a standard bash file.
-
- The following collapsed table contains all the options that can be configured for the sensors collector.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: sensors_sys_dir
- description: The directory the kernel exposes sensor data.
- default_value: "/sys/devices"
- required: false
- - name: sensors_sys_depth
- description: How deep in the tree to check for sensor data.
- default_value: 10
- required: false
- - name: sensors_source_update
- description: If set to 1, the script will overwrite internal script functions with code generated ones.
- default_value: 1
- required: false
- - name: sensors_update_every
- description: The data collection frequency. If unset, will inherit the netdata update frequency.
- default_value: 1
- required: false
- - name: sensors_priority
- description: The charts priority on the dashboard.
- default_value: 90000
- required: false
- - name: sensors_retries
- description: The number of retries to do in case of failure before disabling the collector.
- default_value: 10
- required: false
- examples:
- folding:
- enabled: false
- title: "Config"
- list:
- - name: Set sensors path depth
- description: Set a different sensors path depth
- config: |
- # the directory the kernel keeps sensor data
- #sensors_sys_dir="/sys/devices"
-
- # how deep in the tree to check for sensor data
- sensors_sys_depth=5
-
- # if set to 1, the script will overwrite internal
- # script functions with code generated ones
- # leave to 1, is faster
- #sensors_source_update=1
-
- # the data collection frequency
- # if unset, will inherit the netdata update frequency
- #sensors_update_every=
-
- # the charts priority on the dashboard
- #sensors_priority=90000
-
- # the number of retries to do in case of failure
- # before disabling the module
- #sensors_retries=10
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: sensor chip
- description: "Metrics related to sensor chips. Each chip provides its own set of the following metrics."
- labels: []
- metrics:
- - name: sensors.temp
- description: Temperature
- unit: "Celsius"
- chart_type: line
- dimensions:
- - name: "{filename}"
- - name: sensors.volt
- description: Voltage
- unit: "Volts"
- chart_type: line
- dimensions:
- - name: "{filename}"
- - name: sensors.curr
- description: Current
- unit: "Ampere"
- chart_type: line
- dimensions:
- - name: "{filename}"
- - name: sensors.power
- description: Power
- unit: "Watt"
- chart_type: line
- dimensions:
- - name: "{filename}"
- - name: sensors.fans
- description: Fans Speed
- unit: "Rotations / Minute"
- chart_type: line
- dimensions:
- - name: "{filename}"
- - name: sensors.energy
- description: Energy
- unit: "Joule"
- chart_type: area
- dimensions:
- - name: "{filename}"
- - name: sensors.humidity
- description: Humidity
- unit: "Percent"
- chart_type: line
- dimensions:
- - name: "{filename}"
diff --git a/collectors/checks.plugin/README.md b/collectors/checks.plugin/README.md
deleted file mode 100644
index 801f27752..000000000
--- a/collectors/checks.plugin/README.md
+++ /dev/null
@@ -1,12 +0,0 @@
-<!--
-title: "checks.plugin"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/checks.plugin/README.md"
-sidebar_label: "checks.plugin"
-learn_status: "Unpublished"
--->
-
-# checks.plugin
-
-A debugging plugin (by default it is disabled)
-
-
diff --git a/collectors/cups.plugin/Makefile.am b/collectors/cups.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/cups.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/debugfs.plugin/Makefile.am b/collectors/debugfs.plugin/Makefile.am
deleted file mode 100644
index 02fe3a314..000000000
--- a/collectors/debugfs.plugin/Makefile.am
+++ /dev/null
@@ -1,9 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
-
diff --git a/collectors/debugfs.plugin/README.md b/collectors/debugfs.plugin/README.md
deleted file mode 100644
index a2dc9c0f6..000000000
--- a/collectors/debugfs.plugin/README.md
+++ /dev/null
@@ -1,65 +0,0 @@
-# OS provided metrics (debugfs.plugin)
-
-`debugfs.plugin` gathers metrics from the `/sys/kernel/debug` folder on Linux
-systems. [Debugfs](https://docs.kernel.org/filesystems/debugfs.html) exists as an easy way for kernel developers to
-make information available to user space.
-
-This plugin
-is [external](https://github.com/netdata/netdata/tree/master/collectors#collector-architecture-and-terminology),
-the netdata daemon spawns it as a long-running independent process.
-
-In detail, it collects metrics from:
-
-- `/sys/kernel/debug/extfrag` (Memory fragmentation index for each order and zone).
-- `/sys/kernel/debug/zswap` ([Zswap](https://www.kernel.org/doc/Documentation/vm/zswap.txt) performance statistics).
-
-## Prerequisites
-
-### Permissions
-
-> No user action required.
-
-The debugfs root directory is accessible only to the root user by default. Netdata
-uses [Linux Capabilities](https://man7.org/linux/man-pages/man7/capabilities.7.html) to give the plugin access
-to debugfs. `CAP_DAC_READ_SEARCH` is added automatically during installation. This capability allows bypassing file read
-permission checks and directory read and execute permission checks. If file capabilities are not usable, then the plugin is instead installed with the SUID bit set in permissions so that it runs as root.
-
-## Metrics
-
-| Metric | Scope | Dimensions | Units | Labels |
-|-------------------------------------|:---------:|:---------------------------------------------------------------------------------------:|:------------:|:---------:|
-| mem.fragmentation_index_dma | numa node | order0, order1, order2, order3, order4, order5, order6, order7, order8, order9, order10 | index | numa_node |
-| mem.fragmentation_index_dma32 | numa node | order0, order1, order2, order3, order4, order5, order6, order7, order8, order9, order10 | index | numa_node |
-| mem.fragmentation_index_normal | numa node | order0, order1, order2, order3, order4, order5, order6, order7, order8, order9, order10 | index | numa_node |
-| system.zswap_pool_compression_ratio | | compression_ratio | ratio | |
-| system.zswap_pool_compressed_size | | compressed_size | bytes | |
-| system.zswap_pool_raw_size | | uncompressed_size | bytes | |
-| system.zswap_rejections | | compress_poor, kmemcache_fail, alloc_fail, reclaim_fail | rejections/s | |
-| system.zswap_pool_limit_hit | | limit | events/s | |
-| system.zswap_written_back_raw_bytes | | written_back | bytes/s | |
-| system.zswap_same_filled_raw_size | | same_filled | bytes | |
-| system.zswap_duplicate_entry | | entries | entries/s | |
-
-## Troubleshooting
-
-To troubleshoot issues with the collector, run the `debugfs.plugin` in the terminal. The output
-should give you clues as to why the collector isn't working.
-
-- Navigate to the `plugins.d` directory, usually at `/usr/libexec/netdata/plugins.d/`. If that's not the case on
- your system, open `netdata.conf` and look for the `plugins` setting under `[directories]`.
-
- ```bash
- cd /usr/libexec/netdata/plugins.d/
- ```
-
-- Switch to the `netdata` user.
-
- ```bash
- sudo -u netdata -s
- ```
-
-- Run the `debugfs.plugin` to debug the collector:
-
- ```bash
- ./debugfs.plugin
- ```
diff --git a/collectors/diskspace.plugin/Makefile.am b/collectors/diskspace.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/diskspace.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/diskspace.plugin/metadata.yaml b/collectors/diskspace.plugin/metadata.yaml
deleted file mode 100644
index cb45edf34..000000000
--- a/collectors/diskspace.plugin/metadata.yaml
+++ /dev/null
@@ -1,139 +0,0 @@
-plugin_name: diskspace.plugin
-modules:
- - meta:
- plugin_name: diskspace.plugin
- module_name: diskspace.plugin
- monitored_instance:
- name: Disk space
- link: ""
- categories:
- - data-collection.linux-systems
- icon_filename: "hard-drive.svg"
- related_resources:
- integrations:
- list:
- - plugin_name: ebpf.plugin
- module_name: disk
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - disk
- - I/O
- - space
- - inode
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor Disk space metrics for proficient storage management. Keep track of usage, free space, and error rates to prevent disk space issues."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: "The plugin reads data from `/proc/self/mountinfo` and `/proc/diskstats file`."
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:proc:diskspace]"
- description: "This is netdata main configuration file"
- options:
- description: "You can also specify per mount point `[plugin:proc:diskspace:mountpoint]`"
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 1
- required: false
- - name: remove charts of unmounted disks
- description: Remove chart when a device is unmounted on host.
- default_value: yes
- required: false
- - name: check for new mount points every
- description: Parse proc files frequency.
- default_value: 15
- required: false
- - name: exclude space metrics on paths
- description: Do not show metrics (charts) for listed paths. This option accepts netdata simple pattern.
- default_value: /proc/* /sys/* /var/run/user/* /run/user/* /snap/* /var/lib/docker/*
- required: false
- - name: exclude space metrics on filesystems
- description: Do not show metrics (charts) for listed filesystems. This option accepts netdata simple pattern.
- default_value: "*gvfs *gluster* *s3fs *ipfs *davfs2 *httpfs *sshfs *gdfs *moosefs fusectl autofs"
- required: false
- - name: exclude inode metrics on filesystems
- description: Do not show metrics (charts) for listed filesystems. This option accepts netdata simple pattern.
- default_value: msdosfs msdos vfat overlayfs aufs* *unionfs
- required: false
- - name: space usage for all disks
- description: Define if plugin will show metrics for space usage. When value is set to `auto` plugin will try to access information to display if filesystem or path was not discarded with previous option.
- default_value: auto
- required: false
- - name: inodes usage for all disks
- description: Define if plugin will show metrics for inode usage. When value is set to `auto` plugin will try to access information to display if filesystem or path was not discarded with previous option.
- default_value: auto
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: disk_space_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf
- metric: disk.space
- info: disk ${label:mount_point} space utilization
- os: "linux freebsd"
- - name: disk_inode_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf
- metric: disk.inodes
- info: disk ${label:mount_point} inode utilization
- os: "linux freebsd"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: mount point
- description: ""
- labels:
- - name: mount_point
- description: Path used to mount a filesystem
- - name: filesystem
- description: The filesystem used to format a partition.
- - name: mount_root
- description: Root directory where mount points are present.
- metrics:
- - name: disk.space
- description: Disk Space Usage
- unit: "GiB"
- chart_type: stacked
- dimensions:
- - name: avail
- - name: used
- - name: reserved_for_root
- - name: disk.inodes
- description: Disk Files (inodes) Usage
- unit: "inodes"
- chart_type: stacked
- dimensions:
- - name: avail
- - name: used
- - name: reserved_for_root
diff --git a/collectors/ebpf.plugin/Makefile.am b/collectors/ebpf.plugin/Makefile.am
deleted file mode 100644
index 2d5f92a6b..000000000
--- a/collectors/ebpf.plugin/Makefile.am
+++ /dev/null
@@ -1,42 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-include $(top_srcdir)/build/subst.inc
-SUFFIXES = .in
-
-userebpfconfigdir=$(configdir)/ebpf.d
-
-# Explicitly install directories to avoid permission issues due to umask
-install-exec-local:
- $(INSTALL) -d $(DESTDIR)$(userebpfconfigdir)
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
-
-ebpfconfigdir=$(libconfigdir)/ebpf.d
-dist_libconfig_DATA = \
- ebpf.d.conf \
- $(NULL)
-
-dist_ebpfconfig_DATA = \
- ebpf.d/ebpf_kernel_reject_list.txt \
- ebpf.d/cachestat.conf \
- ebpf.d/dcstat.conf \
- ebpf.d/disk.conf \
- ebpf.d/fd.conf \
- ebpf.d/filesystem.conf \
- ebpf.d/hardirq.conf \
- ebpf.d/mdflush.conf \
- ebpf.d/mount.conf \
- ebpf.d/network.conf \
- ebpf.d/oomkill.conf \
- ebpf.d/process.conf \
- ebpf.d/shm.conf \
- ebpf.d/softirq.conf \
- ebpf.d/sync.conf \
- ebpf.d/swap.conf \
- ebpf.d/vfs.conf \
- $(NULL)
diff --git a/collectors/ebpf.plugin/README.md b/collectors/ebpf.plugin/README.md
deleted file mode 100644
index 06915ea52..000000000
--- a/collectors/ebpf.plugin/README.md
+++ /dev/null
@@ -1,1071 +0,0 @@
-<!--
-title: "Kernel traces/metrics (eBPF) monitoring with Netdata"
-description: "Use Netdata's extended Berkeley Packet Filter (eBPF) collector to monitor kernel-level metrics about yourcomplex applications with per-second granularity."
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/README.md"
-sidebar_label: "Kernel traces/metrics (eBPF)"
-learn_status: "Published"
-learn_topic_type: "References"
-learn_rel_path: "Integrations/Monitor/System metrics"
--->
-
-# Kernel traces/metrics (eBPF) collector
-
-The Netdata Agent provides many [eBPF](https://ebpf.io/what-is-ebpf/) programs to help you troubleshoot and debug how applications interact with the Linux kernel. The `ebpf.plugin` uses [tracepoints, trampoline, and2 kprobes](#how-netdata-collects-data-using-probes-and-tracepoints) to collect a wide array of high value data about the host that would otherwise be impossible to capture.
-
-> ❗ eBPF monitoring only works on Linux systems and with specific Linux kernels, including all kernels newer than `4.11.0`, and all kernels on CentOS 7.6 or later. For kernels older than `4.11.0`, improved support is in active development.
-
-This document provides comprehensive details about the `ebpf.plugin`.
-For hands-on configuration and troubleshooting tips see our [tutorial on troubleshooting apps with eBPF metrics](https://github.com/netdata/netdata/blob/master/docs/guides/troubleshoot/monitor-debug-applications-ebpf.md).
-
-<figure>
- <img src="https://user-images.githubusercontent.com/1153921/74746434-ad6a1e00-5222-11ea-858a-a7882617ae02.png" alt="An example of VFS charts, made possible by the eBPF collector plugin" />
- <figcaption>An example of virtual file system (VFS) charts made possible by the eBPF collector plugin.</figcaption>
-</figure>
-
-## How Netdata collects data using probes and tracepoints
-
-Netdata uses the following features from the Linux kernel to run eBPF programs:
-
-- Tracepoints are hooks to call specific functions. Tracepoints are more stable than `kprobes` and are preferred when
- both options are available.
-- Trampolines are bridges between kernel functions, and BPF programs. Netdata uses them by default whenever available.
-- Kprobes and return probes (`kretprobe`): Probes can insert virtually into any kernel instruction. When eBPF runs in `entry` mode, it attaches only `kprobes` for internal functions monitoring calls and some arguments every time a function is called. The user can also change configuration to use [`return`](#global-configuration-options) mode, and this will allow users to monitor return from these functions and detect possible failures.
-
-In each case, wherever a normal kprobe, kretprobe, or tracepoint would have run its hook function, an eBPF program is run instead, performing various collection logic before letting the kernel continue its normal control flow.
-
-There are more methods to trigger eBPF programs, such as uprobes, but currently are not supported.
-
-## Configuring ebpf.plugin
-
-The eBPF collector is installed and enabled by default on most new installations of the Agent.
-If your Agent is v1.22 or older, you may to enable the collector yourself.
-
-### Enable the eBPF collector
-
-To enable or disable the entire eBPF collector:
-
-1. Navigate to the [Netdata config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
- ```bash
- cd /etc/netdata
- ```
-
-2. Use the [`edit-config`](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#use-edit-config-to-edit-configuration-files) script to edit `netdata.conf`.
-
- ```bash
- ./edit-config netdata.conf
- ```
-
-3. Enable the collector by scrolling down to the `[plugins]` section. Uncomment the line `ebpf` (not
- `ebpf_process`) and set it to `yes`.
-
- ```conf
- [plugins]
- ebpf = yes
- ```
-
-### Configure the eBPF collector
-
-You can configure the eBPF collector's behavior to fine-tune which metrics you receive and [optimize performance]\(#performance opimization).
-
-To edit the `ebpf.d.conf`:
-
-1. Navigate to the [Netdata config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
- ```bash
- cd /etc/netdata
- ```
-2. Use the [`edit-config`](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#use-edit-config-to-edit-configuration-files) script to edit [`ebpf.d.conf`](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/ebpf.d.conf).
-
- ```bash
- ./edit-config ebpf.d.conf
- ```
-
- You can now edit the behavior of the eBPF collector. The following sections describe each configuration option in detail.
-
-### `[global]` configuration options
-
-The `[global]` section defines settings for the whole eBPF collector.
-
-#### eBPF load mode
-
-The collector uses two different eBPF programs. These programs rely on the same functions inside the kernel, but they
-monitor, process, and display different kinds of information.
-
-By default, this plugin uses the `entry` mode. Changing this mode can create significant overhead on your operating
-system, but also offer valuable information if you are developing or debugging software. The `ebpf load mode` option
-accepts the following values:
-
-- `entry`: This is the default mode. In this mode, the eBPF collector only monitors calls for the functions described in
- the sections above, and does not show charts related to errors.
-- `return`: In the `return` mode, the eBPF collector monitors the same kernel functions as `entry`, but also creates new
- charts for the return of these functions, such as errors. Monitoring function returns can help in debugging software,
- such as failing to close file descriptors or creating zombie processes.
-
-#### Integration with `apps.plugin`
-
-The eBPF collector also creates charts for each running application through an integration with the
-[`apps.plugin`](https://github.com/netdata/netdata/blob/master/collectors/apps.plugin/README.md). This integration helps you understand how specific applications
-interact with the Linux kernel.
-
-If you want to enable `apps.plugin` integration, change the "apps" setting to "yes".
-
-```conf
-[global]
- apps = yes
-```
-
-#### Integration with `cgroups.plugin`
-
-The eBPF collector also creates charts for each cgroup through an integration with the
-[`cgroups.plugin`](https://github.com/netdata/netdata/blob/master/collectors/cgroups.plugin/README.md). This integration helps you understand how a specific cgroup
-interacts with the Linux kernel.
-
-The integration with `cgroups.plugin` is disabled by default to avoid creating overhead on your system. If you want to
-_enable_ the integration with `cgroups.plugin`, change the `cgroups` setting to `yes`.
-
-```conf
-[global]
- cgroups = yes
-```
-
-If you do not need to monitor specific metrics for your `cgroups`, you can enable `cgroups` inside
-`ebpf.d.conf`, and then disable the plugin for a specific `thread` by following the steps in the
-[Configuration](#configuring-ebpfplugin) section.
-
-#### Maps per Core
-
-When netdata is running on kernels newer than `4.6` users are allowed to modify how the `ebpf.plugin` creates maps (hash or
-array). When `maps per core` is defined as `yes`, plugin will create a map per core on host, on the other hand,
-when the value is set as `no` only one hash table will be created, this option will use less memory, but it also can
-increase overhead for processes.
-
-#### Collect PID
-
-When one of the previous integrations is enabled, `ebpf.plugin` will use Process Identifier (`PID`) to identify the
-process group for which it needs to plot data.
-
-There are different ways to collect PID, and you can select the way `ebpf.plugin` collects data with the following
-values:
-
-- `real parent`: This is the default mode. Collection will aggregate data for the real parent, the thread that creates
- child threads.
-- `parent`: Parent and real parent are the same when a process starts, but this value can be changed during run time.
-- `all`: This option will store all PIDs that run on the host. Note, this method can be expensive for the host,
- because more memory needs to be allocated and parsed.
-
-The threads that have integration with other collectors have an internal clean up wherein they attach either a
-`trampoline` or a `kprobe` to `release_task` internal function. To avoid `overload` on this function, `ebpf.plugin`
-will only enable these threads integrated with other collectors when the kernel is compiled with
-`CONFIG_DEBUG_INFO_BTF`, unless you enable them manually.
-
-#### Collection period
-
-The plugin uses the option `update every` to define the number of seconds used for eBPF to send data for Netdata. The default value
-is 5 seconds.
-
-#### PID table size
-
-The option `pid table size` defines the maximum number of PIDs stored inside the application hash table. The default value
-is defined according [kernel](https://elixir.bootlin.com/linux/v6.0.19/source/include/linux/threads.h#L28) source code.
-
-#### Integration Dashboard Elements
-
-When an integration is enabled, your dashboard will also show the following cgroups and apps charts using low-level
-Linux metrics:
-
-> Note: The parenthetical accompanying each bulleted item provides the chart name.
-
-- mem
- - Number of processes killed due out of memory. (`oomkills`)
-- process
- - Number of processes created with `do_fork`. (`process_create`)
- - Number of threads created with `do_fork` or `clone (2)`, depending on your system's kernel
- version. (`thread_create`)
- - Number of times that a process called `do_exit`. (`task_exit`)
- - Number of times that a process called `release_task`. (`task_close`)
- - Number of times that an error happened to create thread or process. (`task_error`)
-- swap
- - Number of calls to `swap_readpage`. (`swap_read_call`)
- - Number of calls to `swap_writepage`. (`swap_write_call`)
-- network
- - Number of outbound connections using TCP/IPv4. (`outbound_conn_ipv4`)
- - Number of outbound connections using TCP/IPv6. (`outbound_conn_ipv6`)
- - Number of bytes sent. (`total_bandwidth_sent`)
- - Number of bytes received. (`total_bandwidth_recv`)
- - Number of calls to `tcp_sendmsg`. (`bandwidth_tcp_send`)
- - Number of calls to `tcp_cleanup_rbuf`. (`bandwidth_tcp_recv`)
- - Number of calls to `tcp_retransmit_skb`. (`bandwidth_tcp_retransmit`)
- - Number of calls to `udp_sendmsg`. (`bandwidth_udp_send`)
- - Number of calls to `udp_recvmsg`. (`bandwidth_udp_recv`)
-- file access
- - Number of calls to open files. (`file_open`)
- - Number of calls to open files that returned errors. (`open_error`)
- - Number of files closed. (`file_closed`)
- - Number of calls to close files that returned errors. (`file_error_closed`)
-- vfs
- - Number of calls to `vfs_unlink`. (`file_deleted`)
- - Number of calls to `vfs_write`. (`vfs_write_call`)
- - Number of calls to write a file that returned errors. (`vfs_write_error`)
- - Number of calls to `vfs_read`. (`vfs_read_call`)
- - - Number of calls to read a file that returned errors. (`vfs_read_error`)
- - Number of bytes written with `vfs_write`. (`vfs_write_bytes`)
- - Number of bytes read with `vfs_read`. (`vfs_read_bytes`)
- - Number of calls to `vfs_fsync`. (`vfs_fsync`)
- - Number of calls to sync file that returned errors. (`vfs_fsync_error`)
- - Number of calls to `vfs_open`. (`vfs_open`)
- - Number of calls to open file that returned errors. (`vfs_open_error`)
- - Number of calls to `vfs_create`. (`vfs_create`)
- - Number of calls to open file that returned errors. (`vfs_create_error`)
-- page cache
- - Ratio of pages accessed. (`cachestat_ratio`)
- - Number of modified pages ("dirty"). (`cachestat_dirties`)
- - Number of accessed pages. (`cachestat_hits`)
- - Number of pages brought from disk. (`cachestat_misses`)
-- directory cache
- - Ratio of files available in directory cache. (`dc_hit_ratio`)
- - Number of files accessed. (`dc_reference`)
- - Number of files accessed that were not in cache. (`dc_not_cache`)
- - Number of files not found. (`dc_not_found`)
-- ipc shm
- - Number of calls to `shm_get`. (`shmget_call`)
- - Number of calls to `shm_at`. (`shmat_call`)
- - Number of calls to `shm_dt`. (`shmdt_call`)
- - Number of calls to `shm_ctl`. (`shmctl_call`)
-
-### `[ebpf programs]` configuration options
-
-The eBPF collector enables and runs the following eBPF programs by default:
-
-- `cachestat`: Netdata's eBPF data collector creates charts about the memory page cache. When the integration with
- [`apps.plugin`](https://github.com/netdata/netdata/blob/master/collectors/apps.plugin/README.md) is enabled, this collector creates charts for the whole host _and_
- for each application.
-- `fd` : This eBPF program creates charts that show information about calls to open files.
-- `mount`: This eBPF program creates charts that show calls to syscalls mount(2) and umount(2).
-- `shm`: This eBPF program creates charts that show calls to syscalls shmget(2), shmat(2), shmdt(2) and shmctl(2).
-- `process`: This eBPF program creates charts that show information about process life. When in `return` mode, it also
- creates charts showing errors when these operations are executed.
-- `hardirq`: This eBPF program creates charts that show information about time spent servicing individual hardware
- interrupt requests (hard IRQs).
-- `softirq`: This eBPF program creates charts that show information about time spent servicing individual software
- interrupt requests (soft IRQs).
-- `oomkill`: This eBPF program creates a chart that shows OOM kills for all applications recognized via
- the `apps.plugin` integration. Note that this program will show application charts regardless of whether apps
- integration is turned on or off.
-
-You can also enable the following eBPF programs:
-
-- `dcstat` : This eBPF program creates charts that show information about file access using directory cache. It appends
- `kprobes` for `lookup_fast()` and `d_lookup()` to identify if files are inside directory cache, outside and files are
- not found.
-- `disk` : This eBPF program creates charts that show information about disk latency independent of filesystem.
-- `filesystem` : This eBPF program creates charts that show information about some filesystem latency.
-- `swap` : This eBPF program creates charts that show information about swap access.
-- `mdflush`: This eBPF program creates charts that show information about
-- `sync`: Monitor calls to syscalls sync(2), fsync(2), fdatasync(2), syncfs(2), msync(2), and sync_file_range(2).
-- `socket`: This eBPF program creates charts with information about `TCP` and `UDP` functions, including the
- bandwidth consumed by each.
- multi-device software flushes.
-- `vfs`: This eBPF program creates charts that show information about VFS (Virtual File System) functions.
-
-### Configuring eBPF threads
-
-You can configure each thread of the eBPF data collector. This allows you to overwrite global options defined in `/etc/netdata/ebpf.d.conf` and configure specific options for each thread.
-
-To configure an eBPF thread:
-
-1. Navigate to the [Netdata config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
- ```bash
- cd /etc/netdata
- ```
-2. Use the [`edit-config`](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#use-edit-config-to-edit-configuration-files) script to edit a thread configuration file. The following configuration files are available:
-
- - `network.conf`: Configuration for the [`network` thread](#network-configuration). This config file overwrites the global options and also
- lets you specify which network the eBPF collector monitors.
- - `process.conf`: Configuration for the [`process` thread](#sync-configuration).
- - `cachestat.conf`: Configuration for the `cachestat` thread(#filesystem-configuration).
- - `dcstat.conf`: Configuration for the `dcstat` thread.
- - `disk.conf`: Configuration for the `disk` thread.
- - `fd.conf`: Configuration for the `file descriptor` thread.
- - `filesystem.conf`: Configuration for the `filesystem` thread.
- - `hardirq.conf`: Configuration for the `hardirq` thread.
- - `softirq.conf`: Configuration for the `softirq` thread.
- - `sync.conf`: Configuration for the `sync` thread.
- - `vfs.conf`: Configuration for the `vfs` thread.
-
- ```bash
- ./edit-config FILE.conf
- ```
-
-### Network configuration
-
-The network configuration has specific options to configure which network(s) the eBPF collector monitors. These options
-are divided in the following sections:
-
-#### `[network connections]`
-
-You can configure the information shown with function `ebpf_socket` using the settings in this section.
-
-```conf
-[network connections]
- enabled = yes
- resolve hostname ips = no
- resolve service names = yes
- ports = 1-1024 !145 !domain
- hostnames = !example.com
- ips = !127.0.0.1/8 10.0.0.0/8 172.16.0.0/12 192.168.0.0/16 fc00::/7
-```
-
-When you define a `ports` setting, Netdata will collect network metrics for that specific port. For example, if you
-write `ports = 19999`, Netdata will collect only connections for itself. The `hostnames` setting accepts
-[simple patterns](https://github.com/netdata/netdata/blob/master/libnetdata/simple_pattern/README.md). The `ports`, and `ips` settings accept negation (`!`) to deny
-specific values or asterisk alone to define all values.
-
-In the above example, Netdata will collect metrics for all ports between `1` and `1024`, with the exception of `53` (domain)
-and `145`.
-
-The following options are available:
-
-- `enabled`: Disable network connections monitoring. This can affect directly some funcion output.
-- `resolve hostname ips`: Enable resolving IPs to hostnames. It is disabled by default because it can be too slow.
-- `resolve service names`: Convert destination ports into service names, for example, port `53` protocol `UDP` becomes `domain`.
- all names are read from /etc/services.
-- `ports`: Define the destination ports for Netdata to monitor.
-- `hostnames`: The list of hostnames that can be resolved to an IP address.
-- `ips`: The IP or range of IPs that you want to monitor. You can use IPv4 or IPv6 addresses, use dashes to define a
- range of IPs, or use CIDR values.
-
-By default the traffic table is created using the destination IPs and ports of the sockets. This can be
-changed, so that Netdata uses service names (if possible), by specifying `resolve service name = yes` in the configuration
-section.
-
-#### `[service name]`
-
-Netdata uses the list of services in `/etc/services` to plot network connection charts. If this file does not contain
-the name for a particular service you use in your infrastructure, you will need to add it to the `[service name]`
-section.
-
-For example, Netdata's default port (`19999`) is not listed in `/etc/services`. To associate that port with the Netdata
-service in network connection charts, and thus see the name of the service instead of its port, define it:
-
-```conf
-[service name]
- 19999 = Netdata
-```
-
-### Sync configuration
-
-The sync configuration has specific options to disable monitoring for syscalls. All syscalls are monitored by default.
-
-```conf
-[syscalls]
- sync = yes
- msync = yes
- fsync = yes
- fdatasync = yes
- syncfs = yes
- sync_file_range = yes
-```
-
-### Filesystem configuration
-
-The filesystem configuration has specific options to disable monitoring for filesystems; by default, all filesystems are
-monitored.
-
-```conf
-[filesystem]
- btrfsdist = yes
- ext4dist = yes
- nfsdist = yes
- xfsdist = yes
- zfsdist = yes
-```
-
-The ebpf program `nfsdist` monitors only `nfs` mount points.
-
-## Troubleshooting
-
-If the eBPF collector does not work, you can troubleshoot it by running the `ebpf.plugin` command and investigating its
-output.
-
-```bash
-cd /usr/libexec/netdata/plugins.d/
-sudo su -s /bin/bash ./ebpf.plugin
-```
-
-You can also use `grep` to search the Agent's `error.log` for messages related to eBPF monitoring.
-
-```bash
-grep -i ebpf /var/log/netdata/error.log
-```
-
-### Confirm kernel compatibility
-
-The eBPF collector only works on Linux systems and with specific Linux kernels. We support all kernels more recent than
-`4.11.0`, and all kernels on CentOS 7.6 or later.
-
-You can run our helper script to determine whether your system can support eBPF monitoring. If it returns no output, your system is ready to compile and run the eBPF collector.
-
-```bash
-curl -sSL https://raw.githubusercontent.com/netdata/kernel-collector/master/tools/check-kernel-config.sh | sudo bash
-```
-
-
-If you see a warning about a missing kernel
-configuration (`KPROBES KPROBES_ON_FTRACE HAVE_KPROBES BPF BPF_SYSCALL BPF_JIT`), you will need to recompile your kernel
-to support this configuration. The process of recompiling Linux kernels varies based on your distribution and version.
-Read the documentation for your system's distribution to learn more about the specific workflow for recompiling the
-kernel, ensuring that you set all the necessary
-
-- [Ubuntu](https://wiki.ubuntu.com/Kernel/BuildYourOwnKernel)
-- [Debian](https://kernel-team.pages.debian.net/kernel-handbook/ch-common-tasks.html#s-common-official)
-- [Fedora](https://fedoraproject.org/wiki/Building_a_custom_kernel)
-- [CentOS](https://wiki.centos.org/HowTos/Custom_Kernel)
-- [Arch Linux](https://wiki.archlinux.org/index.php/Kernel/Traditional_compilation)
-- [Slackware](https://docs.slackware.com/howtos:slackware_admin:kernelbuilding)
-
-### Mount `debugfs` and `tracefs`
-
-The eBPF collector also requires both the `tracefs` and `debugfs` filesystems. Try mounting the `tracefs` and `debugfs`
-filesystems using the commands below:
-
-```bash
-sudo mount -t debugfs nodev /sys/kernel/debug
-sudo mount -t tracefs nodev /sys/kernel/tracing
-```
-
-If they are already mounted, you will see an error. You can also configure your system's `/etc/fstab` configuration to
-mount these filesystems on startup. More information can be found in
-the [ftrace documentation](https://www.kernel.org/doc/Documentation/trace/ftrace.txt).
-
-## Charts
-
-The eBPF collector creates charts on different menus, like System Overview, Memory, MD arrays, Disks, Filesystem,
-Mount Points, Networking Stack, systemd Services, and Applications.
-
-The collector stores the actual value inside of its process, but charts only show the difference between the values
-collected in the previous and current seconds.
-
-### System overview
-
-Not all charts within the System Overview menu are enabled by default. Charts that rely on `kprobes` are disabled by default because they add around 100ns overhead for each function call. This is a small number from a human's perspective, but the functions are called many times and create an impact
-on host. See the [configuration](#configuring-ebpfplugin) section for details about how to enable them.
-
-#### Processes
-
-Internally, the Linux kernel treats both processes and threads as `tasks`. To create a thread, the kernel offers a few
-system calls: `fork(2)`, `vfork(2)`, and `clone(2)`. To generate this chart, the eBPF
-collector uses the following `tracepoints` and `kprobe`:
-
-- `sched/sched_process_fork`: Tracepoint called after a call for `fork (2)`, `vfork (2)` and `clone (2)`.
-- `sched/sched_process_exec`: Tracepoint called after a exec-family syscall.
-- `kprobe/kernel_clone`: This is the main [`fork()`](https://elixir.bootlin.com/linux/v5.10/source/kernel/fork.c#L2415)
- routine since kernel `5.10.0` was released.
-- `kprobe/_do_fork`: Like `kernel_clone`, but this was the main function between kernels `4.2.0` and `5.9.16`
-- `kprobe/do_fork`: This was the main function before kernel `4.2.0`.
-
-#### Process Exit
-
-Ending a task requires two steps. The first is a call to the internal function `do_exit`, which notifies the operating
-system that the task is finishing its work. The second step is to release the kernel information with the internal
-function `release_task`. The difference between the two dimensions can help you discover
-[zombie processes](https://en.wikipedia.org/wiki/Zombie_process). To get the metrics, the collector uses:
-
-- `sched/sched_process_exit`: Tracepoint called after a task exits.
-- `kprobe/release_task`: This function is called when a process exits, as the kernel still needs to remove the process
- descriptor.
-
-#### Task error
-
-The functions responsible for ending tasks do not return values, so this chart contains information about failures on
-process and thread creation only.
-
-#### Swap
-
-Inside the swap submenu the eBPF plugin creates the chart `swapcalls`; this chart is displaying when processes are
-calling functions [`swap_readpage` and `swap_writepage`](https://hzliu123.github.io/linux-kernel/Page%20Cache%20in%20Linux%202.6.pdf),
-which are functions responsible for doing IO in swap memory. To collect the exact moment that an access to swap happens,
-the collector attaches `kprobes` for cited functions.
-
-#### Soft IRQ
-
-The following `tracepoints` are used to measure time usage for soft IRQs:
-
-- [`irq/softirq_entry`](https://www.kernel.org/doc/html/latest/core-api/tracepoint.html#c.trace_softirq_entry): Called
- before softirq handler
-- [`irq/softirq_exit`](https://www.kernel.org/doc/html/latest/core-api/tracepoint.html#c.trace_softirq_exit): Called when
- softirq handler returns.
-
-#### Hard IRQ
-
-The following tracepoints are used to measure the latency of servicing a
-hardware interrupt request (hard IRQ).
-
-- [`irq/irq_handler_entry`](https://www.kernel.org/doc/html/latest/core-api/tracepoint.html#c.trace_irq_handler_entry):
- Called immediately before the IRQ action handler.
-- [`irq/irq_handler_exit`](https://www.kernel.org/doc/html/latest/core-api/tracepoint.html#c.trace_irq_handler_exit):
- Called immediately after the IRQ action handler returns.
-- `irq_vectors`: These are traces from `irq_handler_entry` and
- `irq_handler_exit` when an IRQ is handled. The following elements from vector
- are triggered:
- - `irq_vectors/local_timer_entry`
- - `irq_vectors/local_timer_exit`
- - `irq_vectors/reschedule_entry`
- - `irq_vectors/reschedule_exit`
- - `irq_vectors/call_function_entry`
- - `irq_vectors/call_function_exit`
- - `irq_vectors/call_function_single_entry`
- - `irq_vectors/call_function_single_xit`
- - `irq_vectors/irq_work_entry`
- - `irq_vectors/irq_work_exit`
- - `irq_vectors/error_apic_entry`
- - `irq_vectors/error_apic_exit`
- - `irq_vectors/thermal_apic_entry`
- - `irq_vectors/thermal_apic_exit`
- - `irq_vectors/threshold_apic_entry`
- - `irq_vectors/threshold_apic_exit`
- - `irq_vectors/deferred_error_entry`
- - `irq_vectors/deferred_error_exit`
- - `irq_vectors/spurious_apic_entry`
- - `irq_vectors/spurious_apic_exit`
- - `irq_vectors/x86_platform_ipi_entry`
- - `irq_vectors/x86_platform_ipi_exit`
-
-#### IPC shared memory
-
-To monitor shared memory system call counts, Netdata attaches tracing in the following functions:
-
-- `shmget`: Runs when [`shmget`](https://man7.org/linux/man-pages/man2/shmget.2.html) is called.
-- `shmat`: Runs when [`shmat`](https://man7.org/linux/man-pages/man2/shmat.2.html) is called.
-- `shmdt`: Runs when [`shmdt`](https://man7.org/linux/man-pages/man2/shmat.2.html) is called.
-- `shmctl`: Runs when [`shmctl`](https://man7.org/linux/man-pages/man2/shmctl.2.html) is called.
-
-### Memory
-
-In the memory submenu the eBPF plugin creates two submenus **page cache** and **synchronization** with the following
-organization:
-
-- Page Cache
- - Page cache ratio
- - Dirty pages
- - Page cache hits
- - Page cache misses
-- Synchronization
- - File sync
- - Memory map sync
- - File system sync
- - File range sync
-
-#### Page cache hits
-
-When the processor needs to read or write a location in main memory, it checks for a corresponding entry in the page cache.
- If the entry is there, a page cache hit has occurred and the read is from the cache.
-
-A page cache hit is when the page cache is successfully accessed with a read operation. We do not count pages that were
-added relatively recently.
-
-#### Dirty pages
-
-A "dirty page" is a page in the page cache that was modified after being created. Since non-dirty pages in the page cache
- have identical copies in secondary storage (e.g. hard disk drive or solid-state drive), discarding and reusing their space
- is much quicker than paging out application memory, and is often preferred over flushing the dirty pages into secondary storage
- and reusing their space.
-
-On `cachestat_dirties` Netdata demonstrates the number of pages that were modified. This chart shows the number of calls
-to the function `mark_buffer_dirty`.
-
-#### Page cache ratio
-
-When the processor needs to read or write in a specific memory address, it checks for a corresponding entry in the page cache.
-If the processor hits a page cache (`page cache hit`), it reads the entry from the cache. If there is no entry (`page cache miss`),
- the kernel allocates a new entry and copies data from the disk. Netdata calculates the percentage of accessed files that are cached on
- memory. The ratio is calculated counting the accessed cached pages
- (without counting [dirty pages](#dirty-pages) and pages added because of read misses) divided by total access without dirty pages.
-
-> \_\_**\_\_\_\_**<ins>Number of accessed cached pages</ins>\***\*\_\_\*\***<br/>
-> Number of total accessed pages - dirty pages - missed pages
-
-The chart `cachestat_ratio` shows how processes are accessing page cache. In a normal scenario, we expect values around
-100%, which means that the majority of the work on the machine is processed in memory. To calculate the ratio, Netdata
-attaches `kprobes` for kernel functions:
-
-- `add_to_page_cache_lru`: Page addition.
-- `mark_page_accessed`: Access to cache.
-- `account_page_dirtied`: Dirty (modified) pages.
-- `mark_buffer_dirty`: Writes to page cache.
-
-#### Page cache misses
-
-A page cache miss means that a page was not inside memory when the process tried to access it. This chart shows the
-result of the difference for calls between functions `add_to_page_cache_lru` and `account_page_dirtied`.
-
-#### File sync
-
-This chart shows calls to synchronization methods, [`fsync(2)`](https://man7.org/linux/man-pages/man2/fdatasync.2.html)
-and [`fdatasync(2)`](https://man7.org/linux/man-pages/man2/fdatasync.2.html), to transfer all modified page caches
-for the files on disk devices. These calls block until the disk reports that the transfer has been completed. They flush
-data for specific file descriptors.
-
-#### Memory map sync
-
-The chart shows calls to [`msync(2)`](https://man7.org/linux/man-pages/man2/msync.2.html) syscalls. This syscall flushes
-changes to a file that was mapped into memory using [`mmap(2)`](https://man7.org/linux/man-pages/man2/mmap.2.html).
-
-#### File system sync
-
-This chart monitors calls demonstrating commits from filesystem caches to disk. Netdata attaches `tracing` for
-[`sync(2)`](https://man7.org/linux/man-pages/man2/sync.2.html), and [`syncfs(2)`](https://man7.org/linux/man-pages/man2/sync.2.html).
-
-#### File range sync
-
-This chart shows calls to [`sync_file_range(2)`](https://man7.org/linux/man-pages/man2/sync_file_range.2.html) which
-synchronizes file segments with disk.
-
-> Note: This is the most dangerous syscall to synchronize data, according to its manual.
-
-### Multiple Device (MD) arrays
-
-The eBPF plugin shows multi-device flushes happening in real time. This can be used to explain some spikes happening
-in [disk latency](#disk) charts.
-
-By default, MD flush is disabled. To enable it, configure your
-`/etc/netdata/ebpf.d.conf` file as:
-
-```conf
-[global]
- mdflush = yes
-```
-
-#### MD flush
-
-To collect data related to Linux multi-device (MD) flushing, the following kprobe is used:
-
-- `kprobe/md_flush_request`: called whenever a request for flushing multi-device data is made.
-
-### Disk
-
-The eBPF plugin also shows a chart in the Disk section when the `disk` thread is enabled.
-
-#### Disk Latency
-
-This will create the chart `disk_latency_io` for each disk on the host. The following tracepoints are used:
-
-- [`block/block_rq_issue`](https://www.kernel.org/doc/html/latest/core-api/tracepoint.html#c.trace_block_rq_issue):
- IO request operation to a device drive.
-- [`block/block_rq_complete`](https://www.kernel.org/doc/html/latest/core-api/tracepoint.html#c.trace_block_rq_complete):
- IO operation completed by device.
-
-Disk Latency is the single most important metric to focus on when it comes to storage performance, under most circumstances.
-For hard drives, an average latency somewhere between 10 to 20 ms can be considered acceptable. For SSD (Solid State Drives),
-in most cases, workloads experience less than 1 ms latency numbers, but workloads should never reach higher than 3 ms.
-The dimensions refer to time intervals.
-
-### Filesystem
-
-This group has charts demonstrating how applications interact with the Linux kernel to open and close file descriptors.
-It also brings latency charts for several different filesystems.
-
-#### Latency Algorithm
-
-We calculate the difference between the calling and return times, spanning disk I/O, file system operations (lock, I/O),
-run queue latency and all events related to the monitored action.
-
-#### ext4
-
-To measure the latency of executing some actions in an
-[ext4](https://elixir.bootlin.com/linux/latest/source/fs/ext4) filesystem, the
-collector needs to attach `kprobes` and `kretprobes` for each of the following
-functions:
-
-- `ext4_file_read_iter`: Function used to measure read latency.
-- `ext4_file_write_iter`: Function used to measure write latency.
-- `ext4_file_open`: Function used to measure open latency.
-- `ext4_sync_file`: Function used to measure sync latency.
-
-#### ZFS
-
-To measure the latency of executing some actions in a zfs filesystem, the
-collector needs to attach `kprobes` and `kretprobes` for each of the following
-functions:
-
-- `zpl_iter_read`: Function used to measure read latency.
-- `zpl_iter_write`: Function used to measure write latency.
-- `zpl_open`: Function used to measure open latency.
-- `zpl_fsync`: Function used to measure sync latency.
-
-#### XFS
-
-To measure the latency of executing some actions in an
-[xfs](https://elixir.bootlin.com/linux/latest/source/fs/xfs) filesystem, the
-collector needs to attach `kprobes` and `kretprobes` for each of the following
-functions:
-
-- `xfs_file_read_iter`: Function used to measure read latency.
-- `xfs_file_write_iter`: Function used to measure write latency.
-- `xfs_file_open`: Function used to measure open latency.
-- `xfs_file_fsync`: Function used to measure sync latency.
-
-#### NFS
-
-To measure the latency of executing some actions in an
-[nfs](https://elixir.bootlin.com/linux/latest/source/fs/nfs) filesystem, the
-collector needs to attach `kprobes` and `kretprobes` for each of the following
-functions:
-
-- `nfs_file_read`: Function used to measure read latency.
-- `nfs_file_write`: Function used to measure write latency.
-- `nfs_file_open`: Functions used to measure open latency.
-- `nfs4_file_open`: Functions used to measure open latency for NFS v4.
-- `nfs_getattr`: Function used to measure sync latency.
-
-#### btrfs
-
-To measure the latency of executing some actions in a [btrfs](https://elixir.bootlin.com/linux/latest/source/fs/btrfs/file.c)
-filesystem, the collector needs to attach `kprobes` and `kretprobes` for each of the following functions:
-
-> Note: We are listing two functions used to measure `read` latency, but we use either `btrfs_file_read_iter` or
-> `generic_file_read_iter`, depending on kernel version.
-
-- `btrfs_file_read_iter`: Function used to measure read latency since kernel `5.10.0`.
-- `generic_file_read_iter`: Like `btrfs_file_read_iter`, but this function was used before kernel `5.10.0`.
-- `btrfs_file_write_iter`: Function used to write data.
-- `btrfs_file_open`: Function used to open files.
-- `btrfs_sync_file`: Function used to synchronize data to filesystem.
-
-#### File descriptor
-
-To give metrics related to `open` and `close` events, instead of attaching kprobes for each syscall used to do these
-events, the collector attaches `kprobes` for the common function used for syscalls:
-
-- [`do_sys_open`](https://0xax.gitbooks.io/linux-insides/content/SysCall/linux-syscall-5.html): Internal function used to
- open files.
-- [`do_sys_openat2`](https://elixir.bootlin.com/linux/v5.6/source/fs/open.c#L1162):
- Function called from `do_sys_open` since version `5.6.0`.
-- [`close_fd`](https://www.mail-archive.com/linux-kernel@vger.kernel.org/msg2271761.html): Function used to close file
- descriptor since kernel `5.11.0`.
-- `__close_fd`: Function used to close files before version `5.11.0`.
-
-#### File error
-
-This chart shows the number of times some software tried and failed to open or close a file descriptor.
-
-#### VFS
-
-The Linux Virtual File System (VFS) is an abstraction layer on top of a
-concrete filesystem like the ones listed in the parent section, e.g. `ext4`.
-
-In this section we list the mechanism by which we gather VFS data, and what
-charts are consequently created.
-
-##### VFS eBPF Hooks
-
-To measure the latency and total quantity of executing some VFS-level
-functions, ebpf.plugin needs to attach kprobes and kretprobes for each of the
-following functions:
-
-- `vfs_write`: Function used monitoring the number of successful & failed
- filesystem write calls, as well as the total number of written bytes.
-- `vfs_writev`: Same function as `vfs_write` but for vector writes (i.e. a
- single write operation using a group of buffers rather than 1).
-- `vfs_read`: Function used for monitoring the number of successful & failed
- filesystem read calls, as well as the total number of read bytes.
-- `vfs_readv` Same function as `vfs_read` but for vector reads (i.e. a single
- read operation using a group of buffers rather than 1).
-- `vfs_unlink`: Function used for monitoring the number of successful & failed
- filesystem unlink calls.
-- `vfs_fsync`: Function used for monitoring the number of successful & failed
- filesystem fsync calls.
-- `vfs_open`: Function used for monitoring the number of successful & failed
- filesystem open calls.
-- `vfs_create`: Function used for monitoring the number of successful & failed
- filesystem create calls.
-
-##### VFS Deleted objects
-
-This chart monitors calls to `vfs_unlink`. This function is responsible for removing objects from the file system.
-
-##### VFS IO
-
-This chart shows the number of calls to the functions `vfs_read` and `vfs_write`.
-
-##### VFS IO bytes
-
-This chart also monitors `vfs_read` and `vfs_write` but, instead of the number of calls, it shows the total amount of
-bytes read and written with these functions.
-
-The Agent displays the number of bytes written as negative because they are moving down to disk.
-
-##### VFS IO errors
-
-The Agent counts and shows the number of instances where a running program experiences a read or write error.
-
-##### VFS Create
-
-This chart shows the number of calls to `vfs_create`. This function is responsible for creating files.
-
-##### VFS Synchronization
-
-This chart shows the number of calls to `vfs_fsync`. This function is responsible for calling `fsync(2)` or
-`fdatasync(2)` on a file. You can see more details in the Synchronization section.
-
-##### VFS Open
-
-This chart shows the number of calls to `vfs_open`. This function is responsible for opening files.
-
-#### Directory Cache
-
-Metrics for directory cache are collected using kprobe for `lookup_fast`, because we are interested in the number of
-times this function is accessed. On the other hand, for `d_lookup` we are not only interested in the number of times it
-is accessed, but also in possible errors, so we need to attach a `kretprobe`. For this reason, the following is used:
-
-- [`lookup_fast`](https://lwn.net/Articles/649115/): Called to look at data inside the directory cache.
-- [`d_lookup`](https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/fs/dcache.c?id=052b398a43a7de8c68c13e7fa05d6b3d16ce6801#n2223):
- Called when the desired file is not inside the directory cache.
-
-##### Directory Cache Interpretation
-
-When directory cache is showing 100% that means that every accessed file was present in the directory cache.
-If files are not present in the directory cache, they are either not present in the file system or the files were not
-accessed before.
-
-### Mount Points
-
-The following `tracing` are used to collect `mount` & `unmount` call counts:
-
-- [`mount`](https://man7.org/linux/man-pages/man2/mount.2.html): mount filesystem on host.
-- [`umount`](https://man7.org/linux/man-pages/man2/umount.2.html): umount filesystem on host.
-
-### Networking Stack
-
-Netdata monitors socket bandwidth attaching `tracing` for internal functions.
-
-#### TCP outbound connections
-
-This chart demonstrates calls to `tcp_v4_connection` and `tcp_v6_connection` that start connections for IPV4 and IPV6, respectively.
-
-#### TCP inbound connections
-
-This chart demonstrates TCP and UDP connections that the host receives.
-To collect this information, netdata attaches a tracing to `inet_csk_accept`.
-
-#### TCP bandwidth functions
-
-This chart demonstrates calls to functions `tcp_sendmsg`, `tcp_cleanup_rbuf`, and `tcp_close`; these functions are used
-to send & receive data and to close connections when `TCP` protocol is used.
-
-#### TCP bandwidth
-
-This chart demonstrates calls to functions:
-
-- `tcp_sendmsg`: Function responsible to send data for a specified destination.
-- `tcp_cleanup_rbuf`: We use this function instead of `tcp_recvmsg`, because the last one misses `tcp_read_sock` traffic
- and we would also need to add more `tracing` to get the socket and package size.
-- `tcp_close`: Function responsible to close connection.
-
-#### TCP retransmit
-
-This chart demonstrates calls to function `tcp_retransmit` that is responsible for executing TCP retransmission when the
-receiver did not return the packet during the expected time.
-
-#### UDP functions
-
-This chart demonstrates calls to functions `udp_sendmsg` and `udp_recvmsg`, which are responsible for sending &
-receiving data for connections when the `UDP` protocol is used.
-
-#### UDP bandwidth
-
-Like the previous chart, this one also monitors `udp_sendmsg` and `udp_recvmsg`, but instead of showing the number of
-calls, it monitors the number of bytes sent and received.
-
-### Apps
-
-#### OOM Killing
-
-These are tracepoints related to [OOM](https://en.wikipedia.org/wiki/Out_of_memory) killing processes.
-
-- `oom/mark_victim`: Monitors when an oomkill event happens.
-
-## Known issues
-
-### Performance opimization
-
-eBPF monitoring is complex and produces a large volume of metrics. We've discovered scenarios where the eBPF plugin
-significantly increases kernel memory usage by several hundred MB.
-
-When the integration with apps or cgroup is enabled, the eBPF collector allocates memory for each process running. If your
-node is experiencing high memory usage and there is no obvious culprit to be found in the `apps.mem` chart, consider:
-
-- Modify [maps per core](#maps-per-core) to use only one map.
-- Disable [integration with apps](#integration-with-appsplugin).
-- Disable [integration with cgroup](#integration-with-cgroupsplugin).
-
-If with these changes you still suspect eBPF using too much memory, and there is no obvious culprit to be found
-in the `apps.mem` chart, consider testing for high kernel memory usage by [disabling eBPF monitoring](#configuring-ebpfplugin).
-Next, [restart Netdata](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) with
-`sudo systemctl restart netdata` to see if system memory usage (see the `system.ram` chart) has dropped significantly.
-
-Beginning with `v1.31`, kernel memory usage is configurable via the [`pid table size` setting](#pid-table-size)
-in `ebpf.conf`.
-
-The total memory usage is a well known [issue](https://lore.kernel.org/all/167821082315.1693.6957546778534183486.git-patchwork-notify@kernel.org/)
-for eBPF, this is not a bug present in plugin.
-
-### SELinux
-
-When [SELinux](https://www.redhat.com/en/topics/linux/what-is-selinux) is enabled, it may prevent `ebpf.plugin` from
-starting correctly. Check the Agent's `error.log` file for errors like the ones below:
-
-```bash
-2020-06-14 15:32:08: ebpf.plugin ERROR : EBPF PROCESS : Cannot load program: /usr/libexec/netdata/plugins.d/pnetdata_ebpf_process.3.10.0.o (errno 13, Permission denied)
-2020-06-14 15:32:19: netdata ERROR : PLUGINSD[ebpf] : read failed: end of file (errno 9, Bad file descriptor)
-```
-
-You can also check for errors related to `ebpf.plugin` inside `/var/log/audit/audit.log`:
-
-```bash
-type=AVC msg=audit(1586260134.952:97): avc: denied { map_create } for pid=1387 comm="ebpf.pl" scontext=system_u:system_r:unconfined_service_t:s0 tcontext=system_u:system_r:unconfined_service_t:s0 tclass=bpf permissive=0
-type=SYSCALL msg=audit(1586260134.952:97): arch=c000003e syscall=321 success=no exit=-13 a0=0 a1=7ffe6b36f000 a2=70 a3=0 items=0 ppid=1135 pid=1387 auid=4294967295 uid=994 gid=990 euid=0 suid=0 fsuid=0 egid=990 sgid=990 fsgid=990 tty=(none) ses=4294967295 comm="ebpf_proc
-ess.pl" exe="/usr/libexec/netdata/plugins.d/ebpf.plugin" subj=system_u:system_r:unconfined_service_t:s0 key=(null)
-```
-
-If you see similar errors, you will have to adjust SELinux's policies to enable the eBPF collector.
-
-#### Creation of bpf policies
-
-To enable `ebpf.plugin` to run on a distribution with SELinux enabled, it will be necessary to take the following
-actions.
-
-First, stop the Netdata Agent.
-
-```bash
-# systemctl stop netdata
-```
-
-Next, create a policy with the `audit.log` file you examined earlier.
-
-```bash
-# grep ebpf.plugin /var/log/audit/audit.log | audit2allow -M netdata_ebpf
-```
-
-This will create two new files: `netdata_ebpf.te` and `netdata_ebpf.mod`.
-
-Edit the `netdata_ebpf.te` file to change the options `class` and `allow`. You should have the following at the end of
-the `netdata_ebpf.te` file.
-
-```conf
-module netdata_ebpf 1.0;
-require {
- type unconfined_service_t;
- class bpf { map_create map_read map_write prog_load prog_run };
-}
-#============= unconfined_service_t ==============
-allow unconfined_service_t self:bpf { map_create map_read map_write prog_load prog_run };
-```
-
-Then compile your `netdata_ebpf.te` file with the following commands to create a binary that loads the new policies:
-
-```bash
-# checkmodule -M -m -o netdata_ebpf.mod netdata_ebpf.te
-# semodule_package -o netdata_ebpf.pp -m netdata_ebpf.mod
-```
-
-Finally, you can load the new policy and start the Netdata agent again:
-
-```bash
-# semodule -i netdata_ebpf.pp
-# systemctl start netdata
-```
-
-### Linux kernel lockdown
-
-Beginning with [version 5.4](https://www.zdnet.com/article/linux-to-get-kernel-lockdown-feature/), the Linux kernel has
-a feature called "lockdown," which may affect `ebpf.plugin` depending how the kernel was compiled. The following table
-shows how the lockdown module impacts `ebpf.plugin` based on the selected options:
-
-| Enforcing kernel lockdown | Enable lockdown LSM early in init | Default lockdown mode | Can `ebpf.plugin` run with this? |
-| :------------------------ | :-------------------------------- | :-------------------- | :------------------------------- |
-| YES | NO | NO | YES |
-| YES | Yes | None | YES |
-| YES | Yes | Integrity | YES |
-| YES | Yes | Confidentiality | NO |
-
-If you or your distribution compiled the kernel with the last combination, your system cannot load shared libraries
-required to run `ebpf.plugin`.
-
-## Functions
-
-### ebpf_thread
-
-The eBPF plugin has a [function](https://github.com/netdata/netdata/blob/master/docs/cloud/netdata-functions.md) named
-`ebpf_thread` that controls its internal threads and helps to reduce the overhead on host. Using the function you
-can run the plugin with all threads disabled and enable them only when you want to take a look in specific areas.
-
-#### List threads
-
-To list all threads status you can query directly the endpoint function:
-
-`http://localhost:19999/api/v1/function?function=ebpf_thread`
-
-It is also possible to query a specific thread adding keyword `thread` and thread name:
-
-`http://localhost:19999/api/v1/function?function=ebpf_thread%20thread:mount`
-
-#### Enable thread
-
-It is possible to enable a specific thread using the keyword `enable`:
-
-`http://localhost:19999/api/v1/function?function=ebpf_thread%20enable:mount`
-
-this will run thread `mount` during 300 seconds (5 minutes). You can specify a specific period by appending the period
-after the thread name:
-
-`http://localhost:19999/api/v1/function?function=ebpf_thread%20enable:mount:600`
-
-in this example thread `mount` will run during 600 seconds (10 minutes).
-
-#### Disable thread
-
-It is also possible to stop any thread running using the keyword `disable`. For example, to disable `cachestat` you can
-request:
-
-`http://localhost:19999/api/v1/function?function=ebpf_thread%20disable:cachestat`
-
-#### Debugging threads
-
-You can verify the impact of threads on the host by running the
-[ebpf_thread_function.sh](https://github.com/netdata/netdata/blob/master/tests/ebpf/ebpf_thread_function.sh)
-script on your environment.
-
-You can check the results of having threads running on your environment in the Netdata monitoring section on your
-dashboard
-
-<img src="https://github.com/netdata/netdata/assets/49162938/91823573-114c-4c16-b634-cc46f7bb1bcf" alt="Threads running." />
-
-### ebpf_socket
-
-The eBPF plugin has a [function](https://github.com/netdata/netdata/blob/master/docs/cloud/netdata-functions.md) named
-`ebpf_socket` that shows the current status of open sockets on host.
-
-#### Families
-
-The plugin shows by default sockets for IPV4 and IPV6, but it is possible to select a specific family by passing the
-family as an argument:
-
-`http://localhost:19999/api/v1/function?function=ebpf_socket%20family:IPV4`
-
-#### Resolve
-
-The plugin resolves ports to service names by default. You can show the port number by disabling the name resolution:
-
-`http://localhost:19999/api/v1/function?function=ebpf_socket%20resolve:NO`
-
-#### CIDR
-
-The plugin shows connections for all possible destination IPs by default. You can limit the range by specifying the CIDR:
-
-`http://localhost:19999/api/v1/function?function=ebpf_socket%20cidr:192.168.1.0/24`
-
-#### PORT
-
-The plugin shows connections for all possible ports by default. You can limit the range by specifying a port or range
-of ports:
-
-`http://localhost:19999/api/v1/function?function=ebpf_socket%20port:1-1024`
diff --git a/collectors/ebpf.plugin/ebpf.c b/collectors/ebpf.plugin/ebpf.c
deleted file mode 100644
index a8e621643..000000000
--- a/collectors/ebpf.plugin/ebpf.c
+++ /dev/null
@@ -1,4126 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-#include <sys/time.h>
-#include <sys/resource.h>
-#include <ifaddrs.h>
-
-#include "ebpf.h"
-#include "ebpf_socket.h"
-#include "ebpf_unittest.h"
-#include "libnetdata/required_dummies.h"
-
-/*****************************************************************
- *
- * GLOBAL VARIABLES
- *
- *****************************************************************/
-
-char *ebpf_plugin_dir = PLUGINS_DIR;
-static char *ebpf_configured_log_dir = LOG_DIR;
-
-char *ebpf_algorithms[] = {"absolute", "incremental"};
-struct config collector_config = { .first_section = NULL,
- .last_section = NULL,
- .mutex = NETDATA_MUTEX_INITIALIZER,
- .index = { .avl_tree = { .root = NULL, .compar = appconfig_section_compare },
- .rwlock = AVL_LOCK_INITIALIZER } };
-
-int running_on_kernel = 0;
-int ebpf_nprocs;
-int isrh = 0;
-int main_thread_id = 0;
-int process_pid_fd = -1;
-static size_t global_iterations_counter = 1;
-bool publish_internal_metrics = true;
-
-pthread_mutex_t lock;
-pthread_mutex_t ebpf_exit_cleanup;
-pthread_mutex_t collect_data_mutex;
-
-struct netdata_static_thread cgroup_integration_thread = {
- .name = "EBPF CGROUP INT",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
-};
-
-ebpf_module_t ebpf_modules[] = {
- { .info = {.thread_name = "process",
- .config_name = "process",
- .thread_description = NETDATA_EBPF_MODULE_PROCESS_DESC},
- .functions = {.start_routine = ebpf_process_thread,
- .apps_routine = ebpf_process_create_apps_charts,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_LEVEL_REAL_PARENT, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL, .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &process_config,
- .config_file = NETDATA_PROCESS_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_10 |
- NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = NULL, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0 },
- { .info = {.thread_name = "socket",
- .config_name = "socket",
- .thread_description = NETDATA_EBPF_SOCKET_MODULE_DESC},
- .functions = {.start_routine = ebpf_socket_thread,
- .apps_routine = ebpf_socket_create_apps_charts,
- .fnct_routine = ebpf_socket_read_open_connections,
- .fcnt_name = EBPF_FUNCTION_SOCKET,
- .fcnt_desc = EBPF_PLUGIN_SOCKET_FUNCTION_DESCRIPTION,
- .fcnt_thread_chart_name = NULL,
- .fcnt_thread_lifetime_name = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_LEVEL_REAL_PARENT, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL,
- .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &socket_config,
- .config_file = NETDATA_NETWORK_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = socket_targets, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = "cachestat", .config_name = "cachestat", .thread_description = NETDATA_EBPF_CACHESTAT_MODULE_DESC},
- .functions = {.start_routine = ebpf_cachestat_thread,
- .apps_routine = ebpf_cachestat_create_apps_charts,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_LEVEL_REAL_PARENT, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = cachestat_maps, .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &cachestat_config,
- .config_file = NETDATA_CACHESTAT_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18|
- NETDATA_V5_4 | NETDATA_V5_14 | NETDATA_V5_15 | NETDATA_V5_16,
- .load = EBPF_LOAD_LEGACY, .targets = cachestat_targets, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = "sync",
- .config_name = "sync",
- .thread_description = NETDATA_EBPF_SYNC_MODULE_DESC},
- .functions = {.start_routine = ebpf_sync_thread,
- .apps_routine = NULL,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING, .maps = NULL,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_NOT_SET, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &sync_config,
- .config_file = NETDATA_SYNC_CONFIG_FILE,
- // All syscalls have the same kernels
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = sync_targets, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = "dc",
- .config_name = "dc",
- .thread_description = NETDATA_EBPF_DC_MODULE_DESC},
- .functions = {.start_routine = ebpf_dcstat_thread,
- .apps_routine = ebpf_dcstat_create_apps_charts,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_LEVEL_REAL_PARENT, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = dcstat_maps,
- .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &dcstat_config,
- .config_file = NETDATA_DIRECTORY_DCSTAT_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = dc_targets, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = "swap", .config_name = "swap", .thread_description = NETDATA_EBPF_SWAP_MODULE_DESC},
- .functions = {.start_routine = ebpf_swap_thread,
- .apps_routine = ebpf_swap_create_apps_charts,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_LEVEL_REAL_PARENT, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL,
- .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &swap_config,
- .config_file = NETDATA_DIRECTORY_SWAP_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = swap_targets, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = "vfs",
- .config_name = "vfs",
- .thread_description = NETDATA_EBPF_VFS_MODULE_DESC},
- .functions = {.start_routine = ebpf_vfs_thread,
- .apps_routine = ebpf_vfs_create_apps_charts,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_LEVEL_REAL_PARENT, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL,
- .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &vfs_config,
- .config_file = NETDATA_DIRECTORY_VFS_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = vfs_targets, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = "filesystem", .config_name = "filesystem", .thread_description = NETDATA_EBPF_FS_MODULE_DESC},
- .functions = {.start_routine = ebpf_filesystem_thread,
- .apps_routine = NULL,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_NOT_SET, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL, .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &fs_config,
- .config_file = NETDATA_FILESYSTEM_CONFIG_FILE,
- //We are setting kernels as zero, because we load eBPF programs according the kernel running.
- .kernels = 0, .load = EBPF_LOAD_LEGACY, .targets = NULL, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = "disk",
- .config_name = "disk",
- .thread_description = NETDATA_EBPF_DISK_MODULE_DESC},
- .functions = {.start_routine = ebpf_disk_thread,
- .apps_routine = NULL,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_NOT_SET, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL, .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &disk_config,
- .config_file = NETDATA_DISK_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = NULL, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = "mount",
- .config_name = "mount",
- .thread_description = NETDATA_EBPF_MOUNT_MODULE_DESC},
- .functions = {.start_routine = ebpf_mount_thread,
- .apps_routine = NULL,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_NOT_SET, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL, .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &mount_config,
- .config_file = NETDATA_MOUNT_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = mount_targets, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = { .thread_name = "fd",
- .config_name = "fd",
- .thread_description = NETDATA_EBPF_FD_MODULE_DESC},
- .functions = {.start_routine = ebpf_fd_thread,
- .apps_routine = ebpf_fd_create_apps_charts,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_LEVEL_REAL_PARENT, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL,
- .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &fd_config,
- .config_file = NETDATA_FD_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_11 |
- NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = fd_targets, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = { .thread_name = "hardirq",
- .config_name = "hardirq",
- .thread_description = NETDATA_EBPF_HARDIRQ_MODULE_DESC},
- .functions = {.start_routine = ebpf_hardirq_thread,
- .apps_routine = NULL,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_NOT_SET, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL, .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &hardirq_config,
- .config_file = NETDATA_HARDIRQ_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = NULL, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = { .thread_name = "softirq",
- .config_name = "softirq",
- .thread_description = NETDATA_EBPF_SOFTIRQ_MODULE_DESC},
- .functions = {.start_routine = ebpf_softirq_thread,
- .apps_routine = NULL,
- .fnct_routine = NULL },
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_NOT_SET, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL, .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &softirq_config,
- .config_file = NETDATA_SOFTIRQ_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = NULL, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = "oomkill",
- .config_name = "oomkill",
- .thread_description = NETDATA_EBPF_OOMKILL_MODULE_DESC},
- .functions = {.start_routine = ebpf_oomkill_thread,
- .apps_routine = ebpf_oomkill_create_apps_charts,
- .fnct_routine = NULL},.enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_LEVEL_REAL_PARENT, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL,
- .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &oomkill_config,
- .config_file = NETDATA_OOMKILL_CONFIG_FILE,
- .kernels = NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = NULL, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = "shm",
- .config_name = "shm",
- .thread_description = NETDATA_EBPF_SHM_MODULE_DESC},
- .functions = {.start_routine = ebpf_shm_thread,
- .apps_routine = ebpf_shm_create_apps_charts,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_LEVEL_REAL_PARENT, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL,
- .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &shm_config,
- .config_file = NETDATA_DIRECTORY_SHM_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = shm_targets, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = { .thread_name = "mdflush",
- .config_name = "mdflush",
- .thread_description = NETDATA_EBPF_MD_MODULE_DESC},
- .functions = {.start_routine = ebpf_mdflush_thread,
- .apps_routine = NULL,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_NOT_SET, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL, .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = &mdflush_config,
- .config_file = NETDATA_DIRECTORY_MDFLUSH_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = mdflush_targets, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = { .thread_name = "functions",
- .config_name = "functions",
- .thread_description = NETDATA_EBPF_FUNCTIONS_MODULE_DESC},
- .functions = {.start_routine = ebpf_function_thread,
- .apps_routine = NULL,
- .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_RUNNING,
- .update_every = EBPF_DEFAULT_UPDATE_EVERY, .global_charts = 1, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO,
- .apps_level = NETDATA_APPS_NOT_SET, .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0,
- .maps = NULL, .pid_map_size = ND_EBPF_DEFAULT_PID_SIZE, .names = NULL, .cfg = NULL,
- .config_file = NETDATA_DIRECTORY_FUNCTIONS_CONFIG_FILE,
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_14,
- .load = EBPF_LOAD_LEGACY, .targets = NULL, .probe_links = NULL, .objects = NULL,
- .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES, .lifetime = EBPF_DEFAULT_LIFETIME, .running_time = 0},
- { .info = {.thread_name = NULL, .config_name = NULL},
- .functions = {.start_routine = NULL, .apps_routine = NULL, .fnct_routine = NULL},
- .enabled = NETDATA_THREAD_EBPF_NOT_RUNNING, .update_every = EBPF_DEFAULT_UPDATE_EVERY,
- .global_charts = 0, .apps_charts = NETDATA_EBPF_APPS_FLAG_NO, .apps_level = NETDATA_APPS_NOT_SET,
- .cgroup_charts = CONFIG_BOOLEAN_NO, .mode = MODE_ENTRY, .optional = 0, .maps = NULL,
- .pid_map_size = 0, .names = NULL, .cfg = NULL, .kernels = 0, .load = EBPF_LOAD_LEGACY,
- .targets = NULL, .probe_links = NULL, .objects = NULL, .thread = NULL, .maps_per_core = CONFIG_BOOLEAN_YES},
-};
-
-struct netdata_static_thread ebpf_threads[] = {
- {
- .name = "EBPF PROCESS",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF SOCKET",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF CACHESTAT",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF SYNC",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF DCSTAT",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF SWAP",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF VFS",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF FILESYSTEM",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF DISK",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF MOUNT",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF FD",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF HARDIRQ",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF SOFTIRQ",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF OOMKILL",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF SHM",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF MDFLUSH",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 1,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = "EBPF FUNCTIONS",
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
-#ifdef NETDATA_DEV_MODE
- .enabled = 1,
-#else
- .enabled = 0,
-#endif
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
- {
- .name = NULL,
- .config_section = NULL,
- .config_name = NULL,
- .env_name = NULL,
- .enabled = 0,
- .thread = NULL,
- .init_routine = NULL,
- .start_routine = NULL
- },
-};
-
-ebpf_filesystem_partitions_t localfs[] =
- {{.filesystem = "ext4",
- .optional_filesystem = NULL,
- .family = "ext4",
- .objects = NULL,
- .probe_links = NULL,
- .flags = NETDATA_FILESYSTEM_FLAG_NO_PARTITION,
- .enabled = CONFIG_BOOLEAN_YES,
- .addresses = {.function = NULL, .addr = 0},
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4,
- .fs_maps = NULL,
- .fs_obj = NULL,
- .functions = { "ext4_file_read_iter",
- "ext4_file_write_iter",
- "ext4_file_open",
- "ext4_sync_file",
- NULL }},
- {.filesystem = "xfs",
- .optional_filesystem = NULL,
- .family = "xfs",
- .objects = NULL,
- .probe_links = NULL,
- .flags = NETDATA_FILESYSTEM_FLAG_NO_PARTITION,
- .enabled = CONFIG_BOOLEAN_YES,
- .addresses = {.function = NULL, .addr = 0},
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4,
- .fs_maps = NULL,
- .fs_obj = NULL,
- .functions = { "xfs_file_read_iter",
- "xfs_file_write_iter",
- "xfs_file_open",
- "xfs_file_fsync",
- NULL }},
- {.filesystem = "nfs",
- .optional_filesystem = "nfs4",
- .family = "nfs",
- .objects = NULL,
- .probe_links = NULL,
- .flags = NETDATA_FILESYSTEM_ATTR_CHARTS,
- .enabled = CONFIG_BOOLEAN_YES,
- .addresses = {.function = NULL, .addr = 0},
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4,
- .fs_maps = NULL,
- .fs_obj = NULL,
- .functions = { "nfs_file_read",
- "nfs_file_write",
- "nfs_open",
- "nfs_getattr",
- NULL }}, // // "nfs4_file_open" - not present on all kernels
- {.filesystem = "zfs",
- .optional_filesystem = NULL,
- .family = "zfs",
- .objects = NULL,
- .probe_links = NULL,
- .flags = NETDATA_FILESYSTEM_FLAG_NO_PARTITION,
- .enabled = CONFIG_BOOLEAN_YES,
- .addresses = {.function = NULL, .addr = 0},
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4,
- .fs_maps = NULL,
- .fs_obj = NULL,
- .functions = { "zpl_iter_read",
- "zpl_iter_write",
- "zpl_open",
- "zpl_fsync",
- NULL }},
- {.filesystem = "btrfs",
- .optional_filesystem = NULL,
- .family = "btrfs",
- .objects = NULL,
- .probe_links = NULL,
- .flags = NETDATA_FILESYSTEM_FILL_ADDRESS_TABLE,
- .enabled = CONFIG_BOOLEAN_YES,
- .addresses = {.function = "btrfs_file_operations", .addr = 0},
- .kernels = NETDATA_V3_10 | NETDATA_V4_14 | NETDATA_V4_16 | NETDATA_V4_18 | NETDATA_V5_4 | NETDATA_V5_10,
- .fs_maps = NULL,
- .fs_obj = NULL,
- .functions = { "btrfs_file_read_iter",
- "btrfs_file_write_iter",
- "btrfs_file_open",
- "btrfs_sync_file",
- NULL }},
- {.filesystem = NULL,
- .optional_filesystem = NULL,
- .family = NULL,
- .objects = NULL,
- .probe_links = NULL,
- .flags = NETDATA_FILESYSTEM_FLAG_NO_PARTITION,
- .enabled = CONFIG_BOOLEAN_YES,
- .addresses = {.function = NULL, .addr = 0},
- .kernels = 0, .fs_maps = NULL, .fs_obj = NULL}};
-
-ebpf_sync_syscalls_t local_syscalls[] = {
- {.syscall = NETDATA_SYSCALLS_SYNC, .enabled = CONFIG_BOOLEAN_YES, .objects = NULL, .probe_links = NULL,
-#ifdef LIBBPF_MAJOR_VERSION
- .sync_obj = NULL,
-#endif
- .sync_maps = NULL
- },
- {.syscall = NETDATA_SYSCALLS_SYNCFS, .enabled = CONFIG_BOOLEAN_YES, .objects = NULL, .probe_links = NULL,
-#ifdef LIBBPF_MAJOR_VERSION
- .sync_obj = NULL,
-#endif
- .sync_maps = NULL
- },
- {.syscall = NETDATA_SYSCALLS_MSYNC, .enabled = CONFIG_BOOLEAN_YES, .objects = NULL, .probe_links = NULL,
-#ifdef LIBBPF_MAJOR_VERSION
- .sync_obj = NULL,
-#endif
- .sync_maps = NULL
- },
- {.syscall = NETDATA_SYSCALLS_FSYNC, .enabled = CONFIG_BOOLEAN_YES, .objects = NULL, .probe_links = NULL,
-#ifdef LIBBPF_MAJOR_VERSION
- .sync_obj = NULL,
-#endif
- .sync_maps = NULL
- },
- {.syscall = NETDATA_SYSCALLS_FDATASYNC, .enabled = CONFIG_BOOLEAN_YES, .objects = NULL, .probe_links = NULL,
-#ifdef LIBBPF_MAJOR_VERSION
- .sync_obj = NULL,
-#endif
- .sync_maps = NULL
- },
- {.syscall = NETDATA_SYSCALLS_SYNC_FILE_RANGE, .enabled = CONFIG_BOOLEAN_YES, .objects = NULL, .probe_links = NULL,
-#ifdef LIBBPF_MAJOR_VERSION
- .sync_obj = NULL,
-#endif
- .sync_maps = NULL
- },
- {.syscall = NULL, .enabled = CONFIG_BOOLEAN_NO, .objects = NULL, .probe_links = NULL,
-#ifdef LIBBPF_MAJOR_VERSION
- .sync_obj = NULL,
-#endif
- .sync_maps = NULL
- }
-};
-
-
-// Link with cgroup.plugin
-netdata_ebpf_cgroup_shm_t shm_ebpf_cgroup = {NULL, NULL};
-int shm_fd_ebpf_cgroup = -1;
-sem_t *shm_sem_ebpf_cgroup = SEM_FAILED;
-pthread_mutex_t mutex_cgroup_shm;
-
-//Network viewer
-ebpf_network_viewer_options_t network_viewer_opt;
-
-// Statistic
-ebpf_plugin_stats_t plugin_statistics = {.core = 0, .legacy = 0, .running = 0, .threads = 0, .tracepoints = 0,
- .probes = 0, .retprobes = 0, .trampolines = 0, .memlock_kern = 0,
- .hash_tables = 0};
-netdata_ebpf_judy_pid_t ebpf_judy_pid = {.pid_table = NULL, .index = {.JudyLArray = NULL}};
-bool ebpf_plugin_exit = false;
-
-#ifdef LIBBPF_MAJOR_VERSION
-struct btf *default_btf = NULL;
-struct cachestat_bpf *cachestat_bpf_obj = NULL;
-struct dc_bpf *dc_bpf_obj = NULL;
-struct disk_bpf *disk_bpf_obj = NULL;
-struct fd_bpf *fd_bpf_obj = NULL;
-struct hardirq_bpf *hardirq_bpf_obj = NULL;
-struct mdflush_bpf *mdflush_bpf_obj = NULL;
-struct mount_bpf *mount_bpf_obj = NULL;
-struct shm_bpf *shm_bpf_obj = NULL;
-struct socket_bpf *socket_bpf_obj = NULL;
-struct swap_bpf *bpf_obj = NULL;
-struct vfs_bpf *vfs_bpf_obj = NULL;
-#else
-void *default_btf = NULL;
-#endif
-char *btf_path = NULL;
-
-/*****************************************************************
- *
- * FUNCTIONS USED TO MANIPULATE JUDY ARRAY
- *
- *****************************************************************/
-
-/**
- * Hashtable insert unsafe
- *
- * Find or create a value associated to the index
- *
- * @return The lsocket = 0 when new item added to the array otherwise the existing item value is returned in *lsocket
- * we return a pointer to a pointer, so that the caller can put anything needed at the value of the index.
- * The pointer to pointer we return has to be used before any other operation that may change the index (insert/delete).
- *
- */
-void **ebpf_judy_insert_unsafe(PPvoid_t arr, Word_t key)
-{
- JError_t J_Error;
- Pvoid_t *idx = JudyLIns(arr, key, &J_Error);
- if (unlikely(idx == PJERR)) {
- netdata_log_error("Cannot add PID to JudyL, JU_ERRNO_* == %u, ID == %d",
- JU_ERRNO(&J_Error), JU_ERRID(&J_Error));
- }
-
- return idx;
-}
-
-/**
- * Get PID from judy
- *
- * Get a pointer for the `pid` from judy_array;
- *
- * @param judy_array a judy array where PID is the primary key
- * @param pid pid stored.
- */
-netdata_ebpf_judy_pid_stats_t *ebpf_get_pid_from_judy_unsafe(PPvoid_t judy_array, uint32_t pid)
-{
- netdata_ebpf_judy_pid_stats_t **pid_pptr =
- (netdata_ebpf_judy_pid_stats_t **)ebpf_judy_insert_unsafe(judy_array, pid);
- netdata_ebpf_judy_pid_stats_t *pid_ptr = *pid_pptr;
- if (likely(*pid_pptr == NULL)) {
- // a new PID added to the index
- *pid_pptr = aral_mallocz(ebpf_judy_pid.pid_table);
-
- pid_ptr = *pid_pptr;
-
- pid_ptr->cmdline = NULL;
- pid_ptr->socket_stats.JudyLArray = NULL;
- rw_spinlock_init(&pid_ptr->socket_stats.rw_spinlock);
- }
-
- return pid_ptr;
-}
-
-/*****************************************************************
- *
- * FUNCTIONS USED TO ALLOCATE APPS/CGROUP MEMORIES (ARAL)
- *
- *****************************************************************/
-
-/**
- * Allocate PID ARAL
- *
- * Allocate memory using ARAL functions to speed up processing.
- *
- * @param name the internal name used for allocated region.
- * @param size size of each element inside allocated space
- *
- * @return It returns the address on success and NULL otherwise.
- */
-ARAL *ebpf_allocate_pid_aral(char *name, size_t size)
-{
- static size_t max_elements = NETDATA_EBPF_ALLOC_MAX_PID;
- if (max_elements < NETDATA_EBPF_ALLOC_MIN_ELEMENTS) {
- netdata_log_error("Number of elements given is too small, adjusting it for %d", NETDATA_EBPF_ALLOC_MIN_ELEMENTS);
- max_elements = NETDATA_EBPF_ALLOC_MIN_ELEMENTS;
- }
-
- return aral_create(name, size,
- 0, max_elements,
- NULL, NULL, NULL, false, false);
-}
-
-/*****************************************************************
- *
- * FUNCTIONS USED TO CLEAN MEMORY AND OPERATE SYSTEM FILES
- *
- *****************************************************************/
-
-/**
- * Wait to avoid possible coredumps while process is closing.
- */
-static inline void ebpf_check_before2go()
-{
- int i = EBPF_OPTION_ALL_CHARTS;
- usec_t max = USEC_PER_SEC, step = 200000;
- while (i && max) {
- max -= step;
- sleep_usec(step);
- i = 0;
- int j;
- pthread_mutex_lock(&ebpf_exit_cleanup);
- for (j = 0; ebpf_modules[j].info.thread_name != NULL; j++) {
- if (ebpf_modules[j].enabled < NETDATA_THREAD_EBPF_STOPPING)
- i++;
- }
- pthread_mutex_unlock(&ebpf_exit_cleanup);
- }
-
- if (i) {
- netdata_log_error("eBPF cannot unload all threads on time, but it will go away");
- }
-}
-
-/**
- * Close the collector gracefully
- */
-static void ebpf_exit()
-{
-#ifdef LIBBPF_MAJOR_VERSION
- pthread_mutex_lock(&ebpf_exit_cleanup);
- if (default_btf) {
- btf__free(default_btf);
- default_btf = NULL;
- }
- pthread_mutex_unlock(&ebpf_exit_cleanup);
-#endif
-
- char filename[FILENAME_MAX + 1];
- ebpf_pid_file(filename, FILENAME_MAX);
- if (unlink(filename))
- netdata_log_error("Cannot remove PID file %s", filename);
-
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_error("Good bye world! I was PID %d", main_thread_id);
-#endif
- fprintf(stdout, "EXIT\n");
- fflush(stdout);
-
- ebpf_check_before2go();
- pthread_mutex_lock(&mutex_cgroup_shm);
- if (shm_ebpf_cgroup.header) {
- ebpf_unmap_cgroup_shared_memory();
- shm_unlink(NETDATA_SHARED_MEMORY_EBPF_CGROUP_NAME);
- }
- pthread_mutex_unlock(&mutex_cgroup_shm);
-
- exit(0);
-}
-
-/**
- * Unload loegacy code
- *
- * @param objects objects loaded from eBPF programs
- * @param probe_links links from loader
- */
-void ebpf_unload_legacy_code(struct bpf_object *objects, struct bpf_link **probe_links)
-{
- if (!probe_links || !objects)
- return;
-
- struct bpf_program *prog;
- size_t j = 0 ;
- bpf_object__for_each_program(prog, objects) {
- bpf_link__destroy(probe_links[j]);
- j++;
- }
- freez(probe_links);
- if (objects)
- bpf_object__close(objects);
-}
-
-/**
- * Unload Unique maps
- *
- * This function unload all BPF maps from threads using one unique BPF object.
- */
-static void ebpf_unload_unique_maps()
-{
- int i;
- for (i = 0; ebpf_modules[i].info.thread_name; i++) {
- // These threads are cleaned with other functions
- if (i != EBPF_MODULE_SOCKET_IDX)
- continue;
-
- if (ebpf_modules[i].enabled != NETDATA_THREAD_EBPF_STOPPED) {
- if (ebpf_modules[i].enabled != NETDATA_THREAD_EBPF_NOT_RUNNING)
- netdata_log_error("Cannot unload maps for thread %s, because it is not stopped.",
- ebpf_modules[i].info.thread_name);
-
- continue;
- }
-
- if (ebpf_modules[i].load == EBPF_LOAD_LEGACY) {
- ebpf_unload_legacy_code(ebpf_modules[i].objects, ebpf_modules[i].probe_links);
- continue;
- }
-
-#ifdef LIBBPF_MAJOR_VERSION
- if (socket_bpf_obj)
- socket_bpf__destroy(socket_bpf_obj);
-#endif
- }
-}
-
-/**
- * Unload filesystem maps
- *
- * This function unload all BPF maps from filesystem thread.
- */
-static void ebpf_unload_filesystems()
-{
- if (ebpf_modules[EBPF_MODULE_FILESYSTEM_IDX].enabled == NETDATA_THREAD_EBPF_NOT_RUNNING ||
- ebpf_modules[EBPF_MODULE_FILESYSTEM_IDX].enabled < NETDATA_THREAD_EBPF_STOPPING ||
- ebpf_modules[EBPF_MODULE_FILESYSTEM_IDX].load != EBPF_LOAD_LEGACY)
- return;
-
- int i;
- for (i = 0; localfs[i].filesystem != NULL; i++) {
- if (!localfs[i].objects)
- continue;
-
- ebpf_unload_legacy_code(localfs[i].objects, localfs[i].probe_links);
- }
-}
-
-/**
- * Unload sync maps
- *
- * This function unload all BPF maps from sync thread.
- */
-static void ebpf_unload_sync()
-{
- if (ebpf_modules[EBPF_MODULE_SYNC_IDX].enabled == NETDATA_THREAD_EBPF_NOT_RUNNING ||
- ebpf_modules[EBPF_MODULE_SYNC_IDX].enabled < NETDATA_THREAD_EBPF_STOPPING)
- return;
-
- int i;
- for (i = 0; local_syscalls[i].syscall != NULL; i++) {
- if (!local_syscalls[i].enabled)
- continue;
-
-#ifdef LIBBPF_MAJOR_VERSION
- if (local_syscalls[i].sync_obj) {
- sync_bpf__destroy(local_syscalls[i].sync_obj);
- continue;
- }
-#endif
- ebpf_unload_legacy_code(local_syscalls[i].objects, local_syscalls[i].probe_links);
- }
-}
-
-/**
- * Close the collector gracefully
- *
- * @param sig is the signal number used to close the collector
- */
-void ebpf_stop_threads(int sig)
-{
- UNUSED(sig);
- static int only_one = 0;
-
- // Child thread should be closed by itself.
- pthread_mutex_lock(&ebpf_exit_cleanup);
- if (main_thread_id != gettid() || only_one) {
- pthread_mutex_unlock(&ebpf_exit_cleanup);
- return;
- }
- only_one = 1;
- int i;
- for (i = 0; ebpf_modules[i].info.thread_name != NULL; i++) {
- if (ebpf_modules[i].enabled < NETDATA_THREAD_EBPF_STOPPING) {
- netdata_thread_cancel(*ebpf_modules[i].thread->thread);
-#ifdef NETDATA_DEV_MODE
- netdata_log_info("Sending cancel for thread %s", ebpf_modules[i].info.thread_name);
-#endif
- }
- }
- pthread_mutex_unlock(&ebpf_exit_cleanup);
-
- pthread_mutex_lock(&mutex_cgroup_shm);
- netdata_thread_cancel(*cgroup_integration_thread.thread);
-#ifdef NETDATA_DEV_MODE
- netdata_log_info("Sending cancel for thread %s", cgroup_integration_thread.name);
-#endif
- pthread_mutex_unlock(&mutex_cgroup_shm);
-
- ebpf_plugin_exit = true;
-
- ebpf_check_before2go();
-
- pthread_mutex_lock(&ebpf_exit_cleanup);
- ebpf_unload_unique_maps();
- ebpf_unload_filesystems();
- ebpf_unload_sync();
- pthread_mutex_unlock(&ebpf_exit_cleanup);
-
- ebpf_exit();
-}
-
-/*****************************************************************
- *
- * FUNCTIONS TO CREATE CHARTS
- *
- *****************************************************************/
-
-/**
- * Create apps for module
- *
- * Create apps chart that will be used with specific module
- *
- * @param em the module main structure.
- * @param root a pointer for the targets.
- */
-static inline void ebpf_create_apps_for_module(ebpf_module_t *em, struct ebpf_target *root) {
- if (em->enabled < NETDATA_THREAD_EBPF_STOPPING && em->apps_charts && em->functions.apps_routine)
- em->functions.apps_routine(em, root);
-}
-
-/**
- * Create apps charts
- *
- * Call ebpf_create_chart to create the charts on apps submenu.
- *
- * @param root a pointer for the targets.
- */
-static void ebpf_create_apps_charts(struct ebpf_target *root)
-{
- if (unlikely(!ebpf_all_pids))
- return;
-
- struct ebpf_target *w;
- int newly_added = 0;
-
- for (w = root; w; w = w->next) {
- if (w->target)
- continue;
-
- if (unlikely(w->processes && (debug_enabled || w->debug_enabled))) {
- struct ebpf_pid_on_target *pid_on_target;
-
- fprintf(
- stderr, "ebpf.plugin: target '%s' has aggregated %u process%s:", w->name, w->processes,
- (w->processes == 1) ? "" : "es");
-
- for (pid_on_target = w->root_pid; pid_on_target; pid_on_target = pid_on_target->next) {
- fprintf(stderr, " %d", pid_on_target->pid);
- }
-
- fputc('\n', stderr);
- }
-
- if (!w->exposed && w->processes) {
- newly_added++;
- w->exposed = 1;
- if (debug_enabled || w->debug_enabled)
- debug_log_int("%s just added - regenerating charts.", w->name);
- }
- }
-
- int i;
- if (!newly_added) {
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX ; i++) {
- ebpf_module_t *current = &ebpf_modules[i];
- if (current->apps_charts & NETDATA_EBPF_APPS_FLAG_CHART_CREATED)
- continue;
-
- ebpf_create_apps_for_module(current, root);
- }
- return;
- }
-
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX ; i++) {
- ebpf_module_t *current = &ebpf_modules[i];
- ebpf_create_apps_for_module(current, root);
- }
-}
-
-/**
- * Get a value from a structure.
- *
- * @param basis it is the first address of the structure
- * @param offset it is the offset of the data you want to access.
- * @return
- */
-collected_number get_value_from_structure(char *basis, size_t offset)
-{
- collected_number *value = (collected_number *)(basis + offset);
-
- collected_number ret = (collected_number)llabs(*value);
- // this reset is necessary to avoid keep a constant value while processing is not executing a task
- *value = 0;
-
- return ret;
-}
-
-/**
- * Write set command on standard output
- *
- * @param dim the dimension name
- * @param value the value for the dimension
- */
-void write_chart_dimension(char *dim, long long value)
-{
- printf("SET %s = %lld\n", dim, value);
-}
-
-/**
- * Call the necessary functions to create a chart.
- *
- * @param name the chart name
- * @param family the chart family
- * @param move the pointer with the values that will be published
- * @param end the number of values that will be written on standard output
- *
- * @return It returns a variable that maps the charts that did not have zero values.
- */
-void write_count_chart(char *name, char *family, netdata_publish_syscall_t *move, uint32_t end)
-{
- ebpf_write_begin_chart(family, name, "");
-
- uint32_t i = 0;
- while (move && i < end) {
- write_chart_dimension(move->name, move->ncall);
-
- move = move->next;
- i++;
- }
-
- ebpf_write_end_chart();
-}
-
-/**
- * Call the necessary functions to create a chart.
- *
- * @param name the chart name
- * @param family the chart family
- * @param move the pointer with the values that will be published
- * @param end the number of values that will be written on standard output
- */
-void write_err_chart(char *name, char *family, netdata_publish_syscall_t *move, int end)
-{
- ebpf_write_begin_chart(family, name, "");
-
- int i = 0;
- while (move && i < end) {
- write_chart_dimension(move->name, move->nerr);
-
- move = move->next;
- i++;
- }
-
- ebpf_write_end_chart();
-}
-
-/**
- * Write charts
- *
- * Write the current information to publish the charts.
- *
- * @param family chart family
- * @param chart chart id
- * @param dim dimension name
- * @param v1 value.
- */
-void ebpf_one_dimension_write_charts(char *family, char *chart, char *dim, long long v1)
-{
- ebpf_write_begin_chart(family, chart, "");
-
- write_chart_dimension(dim, v1);
-
- ebpf_write_end_chart();
-}
-
-/**
- * Call the necessary functions to create a chart.
- *
- * @param chart the chart name
- * @param family the chart family
- * @param dwrite the dimension name
- * @param vwrite the value for previous dimension
- * @param dread the dimension name
- * @param vread the value for previous dimension
- *
- * @return It returns a variable that maps the charts that did not have zero values.
- */
-void write_io_chart(char *chart, char *family, char *dwrite, long long vwrite, char *dread, long long vread)
-{
- ebpf_write_begin_chart(family, chart, "");
-
- write_chart_dimension(dwrite, vwrite);
- write_chart_dimension(dread, vread);
-
- ebpf_write_end_chart();
-}
-
-/**
- * Write chart cmd on standard output
- *
- * @param type chart type
- * @param id chart id (the apps group name).
- * @param suffix suffix to differentiate charts
- * @param title chart title
- * @param units units label
- * @param family group name used to attach the chart on dashboard
- * @param charttype chart type
- * @param context chart context
- * @param order chart order
- * @param update_every update interval used by plugin
- * @param module chart module name, this is the eBPF thread.
- */
-void ebpf_write_chart_cmd(char *type, char *id, char *suffix, char *title, char *units, char *family,
- char *charttype, char *context, int order, int update_every, char *module)
-{
- printf("CHART %s.%s%s '' '%s' '%s' '%s' '%s' '%s' %d %d '' 'ebpf.plugin' '%s'\n",
- type,
- id,
- suffix,
- title,
- units,
- (family)?family:"",
- (context)?context:"",
- (charttype)?charttype:"",
- order,
- update_every,
- module);
-}
-
-/**
- * Write chart cmd on standard output
- *
- * @param type chart type
- * @param id chart id
- * @param suffix add suffix to obsolete charts.
- * @param title chart title
- * @param units units label
- * @param family group name used to attach the chart on dashboard
- * @param charttype chart type
- * @param context chart context
- * @param order chart order
- * @param update_every value to overwrite the update frequency set by the server.
- */
-void ebpf_write_chart_obsolete(char *type, char *id, char *suffix, char *title, char *units, char *family,
- char *charttype, char *context, int order, int update_every)
-{
- printf("CHART %s.%s%s '' '%s' '%s' '%s' '%s' '%s' %d %d 'obsolete'\n",
- type,
- id,
- suffix,
- title,
- units,
- (family)?family:"",
- (context)?context:"",
- (charttype)?charttype:"",
- order,
- update_every);
-}
-
-/**
- * Write the dimension command on standard output
- *
- * @param name the dimension name
- * @param id the dimension id
- * @param algo the dimension algorithm
- */
-void ebpf_write_global_dimension(char *name, char *id, char *algorithm)
-{
- printf("DIMENSION %s %s %s 1 1\n", name, id, algorithm);
-}
-
-/**
- * Call ebpf_write_global_dimension to create the dimensions for a specific chart
- *
- * @param ptr a pointer to a structure of the type netdata_publish_syscall_t
- * @param end the number of dimensions for the structure ptr
- */
-void ebpf_create_global_dimension(void *ptr, int end)
-{
- netdata_publish_syscall_t *move = ptr;
-
- int i = 0;
- while (move && i < end) {
- ebpf_write_global_dimension(move->name, move->dimension, move->algorithm);
-
- move = move->next;
- i++;
- }
-}
-
-/**
- * Call write_chart_cmd to create the charts
- *
- * @param type chart type
- * @param id chart id
- * @param title chart title
- * @param units axis label
- * @param family group name used to attach the chart on dashboard
- * @param context chart context
- * @param charttype chart type
- * @param order order number of the specified chart
- * @param ncd a pointer to a function called to create dimensions
- * @param move a pointer for a structure that has the dimensions
- * @param end number of dimensions for the chart created
- * @param update_every update interval used with chart.
- * @param module chart module name, this is the eBPF thread.
- */
-void ebpf_create_chart(char *type,
- char *id,
- char *title,
- char *units,
- char *family,
- char *context,
- char *charttype,
- int order,
- void (*ncd)(void *, int),
- void *move,
- int end,
- int update_every,
- char *module)
-{
- ebpf_write_chart_cmd(type, id, "", title, units, family, charttype, context, order, update_every, module);
-
- if (ncd) {
- ncd(move, end);
- }
-}
-
-/**
- * Call the necessary functions to create a name.
- *
- * @param family family name
- * @param name chart name
- * @param hist0 histogram values
- * @param dimensions dimension values.
- * @param end number of bins that will be sent to Netdata.
- *
- * @return It returns a variable that maps the charts that did not have zero values.
- */
-void write_histogram_chart(char *family, char *name, const netdata_idx_t *hist, char **dimensions, uint32_t end)
-{
- ebpf_write_begin_chart(family, name, "");
-
- uint32_t i;
- for (i = 0; i < end; i++) {
- write_chart_dimension(dimensions[i], (long long) hist[i]);
- }
-
- ebpf_write_end_chart();
-
- fflush(stdout);
-}
-
-/**
- * ARAL Charts
- *
- * Add chart to monitor ARAL usage
- * Caller must call this function with mutex locked.
- *
- * @param name the name used to create aral
- * @param em a pointer to the structure with the default values.
- */
-int ebpf_statistic_create_aral_chart(char *name, ebpf_module_t *em)
-{
- static int priority = NETATA_EBPF_ORDER_STAT_ARAL_BEGIN;
- char *mem = { NETDATA_EBPF_STAT_DIMENSION_MEMORY };
- char *aral = { NETDATA_EBPF_STAT_DIMENSION_ARAL };
-
- snprintfz(em->memory_usage, NETDATA_EBPF_CHART_MEM_LENGTH -1, "aral_%s_size", name);
- snprintfz(em->memory_allocations, NETDATA_EBPF_CHART_MEM_LENGTH -1, "aral_%s_alloc", name);
-
- ebpf_write_chart_cmd(NETDATA_MONITORING_FAMILY,
- em->memory_usage,
- "",
- "Bytes allocated for ARAL.",
- "bytes",
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- "netdata.ebpf_aral_stat_size",
- priority++,
- em->update_every,
- NETDATA_EBPF_MODULE_NAME_PROCESS);
-
- ebpf_write_global_dimension(mem,
- mem,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
-
- ebpf_write_chart_cmd(NETDATA_MONITORING_FAMILY,
- em->memory_allocations,
- "",
- "Calls to allocate memory.",
- "calls",
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- "netdata.ebpf_aral_stat_alloc",
- priority++,
- em->update_every,
- NETDATA_EBPF_MODULE_NAME_PROCESS);
-
- ebpf_write_global_dimension(aral,
- aral,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
-
- return priority - 2;
-}
-
-/**
- * ARAL Charts
- *
- * Add chart to monitor ARAL usage
- * Caller must call this function with mutex locked.
- *
- * @param em a pointer to the structure with the default values.
- * @param prio the initial priority used to disable charts.
- */
-void ebpf_statistic_obsolete_aral_chart(ebpf_module_t *em, int prio)
-{
- ebpf_write_chart_obsolete(NETDATA_MONITORING_FAMILY,
- em->memory_allocations,
- "",
- "Calls to allocate memory.",
- "calls",
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- "netdata.ebpf_aral_stat_alloc",
- prio++,
- em->update_every);
-
- ebpf_write_chart_obsolete(NETDATA_MONITORING_FAMILY,
- em->memory_allocations,
- "",
- "Calls to allocate memory.",
- "calls",
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- "netdata.ebpf_aral_stat_alloc",
- prio++,
- em->update_every);
-}
-
-/**
- * Send data from aral chart
- *
- * Send data for eBPF plugin
- *
- * @param memory a pointer to the allocated address
- * @param em a pointer to the structure with the default values.
- */
-void ebpf_send_data_aral_chart(ARAL *memory, ebpf_module_t *em)
-{
- char *mem = { NETDATA_EBPF_STAT_DIMENSION_MEMORY };
- char *aral = { NETDATA_EBPF_STAT_DIMENSION_ARAL };
-
- struct aral_statistics *stats = aral_statistics(memory);
-
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, em->memory_usage, "");
- write_chart_dimension(mem, (long long)stats->structures.allocated_bytes);
- ebpf_write_end_chart();
-
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, em->memory_allocations, "");
- write_chart_dimension(aral, (long long)stats->structures.allocations);
- ebpf_write_end_chart();
-}
-
-/*****************************************************************
- *
- * FUNCTIONS TO READ GLOBAL HASH TABLES
- *
- *****************************************************************/
-
-/**
- * Read Global Table Stats
- *
- * Read data from specified table (map_fd) using array allocated inside thread(values) and storing
- * them in stats vector starting from the first position.
- *
- * For PID tables is recommended to use a function to parse the specific data.
- *
- * @param stats vector used to store data
- * @param values helper to read data from hash tables.
- * @param map_fd table that has data
- * @param maps_per_core Is necessary to read data from all cores?
- * @param begin initial value to query hash table
- * @param end last value that will not be used.
- */
-void ebpf_read_global_table_stats(netdata_idx_t *stats,
- netdata_idx_t *values,
- int map_fd,
- int maps_per_core,
- uint32_t begin,
- uint32_t end)
-{
- uint32_t idx, order;
-
- for (idx = begin, order = 0; idx < end; idx++, order++) {
- if (!bpf_map_lookup_elem(map_fd, &idx, values)) {
- int i;
- int before = (maps_per_core) ? ebpf_nprocs: 1;
- netdata_idx_t total = 0;
- for (i = 0; i < before; i++)
- total += values[i];
-
- stats[order] = total;
- }
- }
-}
-
-/*****************************************************************
- *
- * FUNCTIONS USED WITH SOCKET
- *
- *****************************************************************/
-
-/**
- * Netmask
- *
- * Copied from iprange (https://github.com/firehol/iprange/blob/master/iprange.h)
- *
- * @param prefix create the netmask based in the CIDR value.
- *
- * @return
- */
-static inline in_addr_t ebpf_netmask(int prefix) {
-
- if (prefix == 0)
- return (~((in_addr_t) - 1));
- else
- return (in_addr_t)(~((1 << (32 - prefix)) - 1));
-
-}
-
-/**
- * Broadcast
- *
- * Copied from iprange (https://github.com/firehol/iprange/blob/master/iprange.h)
- *
- * @param addr is the ip address
- * @param prefix is the CIDR value.
- *
- * @return It returns the last address of the range
- */
-static inline in_addr_t ebpf_broadcast(in_addr_t addr, int prefix)
-{
- return (addr | ~ebpf_netmask(prefix));
-}
-
-/**
- * Network
- *
- * Copied from iprange (https://github.com/firehol/iprange/blob/master/iprange.h)
- *
- * @param addr is the ip address
- * @param prefix is the CIDR value.
- *
- * @return It returns the first address of the range.
- */
-static inline in_addr_t ebpf_ipv4_network(in_addr_t addr, int prefix)
-{
- return (addr & ebpf_netmask(prefix));
-}
-
-/**
- * Calculate ipv6 first address
- *
- * @param out the address to store the first address.
- * @param in the address used to do the math.
- * @param prefix number of bits used to calculate the address
- */
-static void get_ipv6_first_addr(union netdata_ip_t *out, union netdata_ip_t *in, uint64_t prefix)
-{
- uint64_t mask,tmp;
- uint64_t ret[2];
-
- memcpy(ret, in->addr32, sizeof(union netdata_ip_t));
-
- if (prefix == 128) {
- memcpy(out->addr32, in->addr32, sizeof(union netdata_ip_t));
- return;
- } else if (!prefix) {
- ret[0] = ret[1] = 0;
- memcpy(out->addr32, ret, sizeof(union netdata_ip_t));
- return;
- } else if (prefix <= 64) {
- ret[1] = 0ULL;
-
- tmp = be64toh(ret[0]);
- mask = 0xFFFFFFFFFFFFFFFFULL << (64 - prefix);
- tmp &= mask;
- ret[0] = htobe64(tmp);
- } else {
- mask = 0xFFFFFFFFFFFFFFFFULL << (128 - prefix);
- tmp = be64toh(ret[1]);
- tmp &= mask;
- ret[1] = htobe64(tmp);
- }
-
- memcpy(out->addr32, ret, sizeof(union netdata_ip_t));
-}
-
-/**
- * Get IPV6 Last Address
- *
- * @param out the address to store the last address.
- * @param in the address used to do the math.
- * @param prefix number of bits used to calculate the address
- */
-static void get_ipv6_last_addr(union netdata_ip_t *out, union netdata_ip_t *in, uint64_t prefix)
-{
- uint64_t mask,tmp;
- uint64_t ret[2];
- memcpy(ret, in->addr32, sizeof(union netdata_ip_t));
-
- if (prefix == 128) {
- memcpy(out->addr32, in->addr32, sizeof(union netdata_ip_t));
- return;
- } else if (!prefix) {
- ret[0] = ret[1] = 0xFFFFFFFFFFFFFFFF;
- memcpy(out->addr32, ret, sizeof(union netdata_ip_t));
- return;
- } else if (prefix <= 64) {
- ret[1] = 0xFFFFFFFFFFFFFFFFULL;
-
- tmp = be64toh(ret[0]);
- mask = 0xFFFFFFFFFFFFFFFFULL << (64 - prefix);
- tmp |= ~mask;
- ret[0] = htobe64(tmp);
- } else {
- mask = 0xFFFFFFFFFFFFFFFFULL << (128 - prefix);
- tmp = be64toh(ret[1]);
- tmp |= ~mask;
- ret[1] = htobe64(tmp);
- }
-
- memcpy(out->addr32, ret, sizeof(union netdata_ip_t));
-}
-
-/**
- * IP to network long
- *
- * @param dst the vector to store the result
- * @param ip the source ip given by our users.
- * @param domain the ip domain (IPV4 or IPV6)
- * @param source the original string
- *
- * @return it returns 0 on success and -1 otherwise.
- */
-static inline int ebpf_ip2nl(uint8_t *dst, char *ip, int domain, char *source)
-{
- if (inet_pton(domain, ip, dst) <= 0) {
- netdata_log_error("The address specified (%s) is invalid ", source);
- return -1;
- }
-
- return 0;
-}
-
-/**
- * Clean port Structure
- *
- * Clean the allocated list.
- *
- * @param clean the list that will be cleaned
- */
-void ebpf_clean_port_structure(ebpf_network_viewer_port_list_t **clean)
-{
- ebpf_network_viewer_port_list_t *move = *clean;
- while (move) {
- ebpf_network_viewer_port_list_t *next = move->next;
- freez(move->value);
- freez(move);
-
- move = next;
- }
- *clean = NULL;
-}
-
-/**
- * Clean IP structure
- *
- * Clean the allocated list.
- *
- * @param clean the list that will be cleaned
- */
-void ebpf_clean_ip_structure(ebpf_network_viewer_ip_list_t **clean)
-{
- ebpf_network_viewer_ip_list_t *move = *clean;
- while (move) {
- ebpf_network_viewer_ip_list_t *next = move->next;
- freez(move->value);
- freez(move);
-
- move = next;
- }
- *clean = NULL;
-}
-
-/**
- * Parse IP List
- *
- * Parse IP list and link it.
- *
- * @param out a pointer to store the link list
- * @param ip the value given as parameter
- */
-static void ebpf_parse_ip_list_unsafe(void **out, char *ip)
-{
- ebpf_network_viewer_ip_list_t **list = (ebpf_network_viewer_ip_list_t **)out;
-
- char *ipdup = strdupz(ip);
- union netdata_ip_t first = { };
- union netdata_ip_t last = { };
- char *is_ipv6;
- if (*ip == '*' && *(ip+1) == '\0') {
- memset(first.addr8, 0, sizeof(first.addr8));
- memset(last.addr8, 0xFF, sizeof(last.addr8));
-
- is_ipv6 = ip;
-
- ebpf_clean_ip_structure(list);
- goto storethisip;
- }
-
- char *end = ip;
- // Move while I cannot find a separator
- while (*end && *end != '/' && *end != '-') end++;
-
- // We will use only the classic IPV6 for while, but we could consider the base 85 in a near future
- // https://tools.ietf.org/html/rfc1924
- is_ipv6 = strchr(ip, ':');
-
- int select;
- if (*end && !is_ipv6) { // IPV4 range
- select = (*end == '/') ? 0 : 1;
- *end++ = '\0';
- if (*end == '!') {
- netdata_log_info("The exclusion cannot be in the second part of the range %s, it will be ignored.", ipdup);
- goto cleanipdup;
- }
-
- if (!select) { // CIDR
- select = ebpf_ip2nl(first.addr8, ip, AF_INET, ipdup);
- if (select)
- goto cleanipdup;
-
- select = (int) str2i(end);
- if (select < NETDATA_MINIMUM_IPV4_CIDR || select > NETDATA_MAXIMUM_IPV4_CIDR) {
- netdata_log_info("The specified CIDR %s is not valid, the IP %s will be ignored.", end, ip);
- goto cleanipdup;
- }
-
- last.addr32[0] = htonl(ebpf_broadcast(ntohl(first.addr32[0]), select));
- // This was added to remove
- // https://app.codacy.com/manual/netdata/netdata/pullRequest?prid=5810941&bid=19021977
- UNUSED(last.addr32[0]);
-
- uint32_t ipv4_test = htonl(ebpf_ipv4_network(ntohl(first.addr32[0]), select));
- if (first.addr32[0] != ipv4_test) {
- first.addr32[0] = ipv4_test;
- struct in_addr ipv4_convert;
- ipv4_convert.s_addr = ipv4_test;
- char ipv4_msg[INET_ADDRSTRLEN];
- if(inet_ntop(AF_INET, &ipv4_convert, ipv4_msg, INET_ADDRSTRLEN))
- netdata_log_info("The network value of CIDR %s was updated for %s .", ipdup, ipv4_msg);
- }
- } else { // Range
- select = ebpf_ip2nl(first.addr8, ip, AF_INET, ipdup);
- if (select)
- goto cleanipdup;
-
- select = ebpf_ip2nl(last.addr8, end, AF_INET, ipdup);
- if (select)
- goto cleanipdup;
- }
-
- if (htonl(first.addr32[0]) > htonl(last.addr32[0])) {
- netdata_log_info("The specified range %s is invalid, the second address is smallest than the first, it will be ignored.",
- ipdup);
- goto cleanipdup;
- }
- } else if (is_ipv6) { // IPV6
- if (!*end) { // Unique
- select = ebpf_ip2nl(first.addr8, ip, AF_INET6, ipdup);
- if (select)
- goto cleanipdup;
-
- memcpy(last.addr8, first.addr8, sizeof(first.addr8));
- } else if (*end == '-') {
- *end++ = 0x00;
- if (*end == '!') {
- netdata_log_info("The exclusion cannot be in the second part of the range %s, it will be ignored.", ipdup);
- goto cleanipdup;
- }
-
- select = ebpf_ip2nl(first.addr8, ip, AF_INET6, ipdup);
- if (select)
- goto cleanipdup;
-
- select = ebpf_ip2nl(last.addr8, end, AF_INET6, ipdup);
- if (select)
- goto cleanipdup;
- } else { // CIDR
- *end++ = 0x00;
- if (*end == '!') {
- netdata_log_info("The exclusion cannot be in the second part of the range %s, it will be ignored.", ipdup);
- goto cleanipdup;
- }
-
- select = str2i(end);
- if (select < 0 || select > 128) {
- netdata_log_info("The CIDR %s is not valid, the address %s will be ignored.", end, ip);
- goto cleanipdup;
- }
-
- uint64_t prefix = (uint64_t)select;
- select = ebpf_ip2nl(first.addr8, ip, AF_INET6, ipdup);
- if (select)
- goto cleanipdup;
-
- get_ipv6_last_addr(&last, &first, prefix);
-
- union netdata_ip_t ipv6_test;
- get_ipv6_first_addr(&ipv6_test, &first, prefix);
-
- if (memcmp(first.addr8, ipv6_test.addr8, sizeof(union netdata_ip_t)) != 0) {
- memcpy(first.addr8, ipv6_test.addr8, sizeof(union netdata_ip_t));
-
- struct in6_addr ipv6_convert;
- memcpy(ipv6_convert.s6_addr, ipv6_test.addr8, sizeof(union netdata_ip_t));
-
- char ipv6_msg[INET6_ADDRSTRLEN];
- if(inet_ntop(AF_INET6, &ipv6_convert, ipv6_msg, INET6_ADDRSTRLEN))
- netdata_log_info("The network value of CIDR %s was updated for %s .", ipdup, ipv6_msg);
- }
- }
-
- if ((be64toh(*(uint64_t *)&first.addr32[2]) > be64toh(*(uint64_t *)&last.addr32[2]) &&
- !memcmp(first.addr32, last.addr32, 2*sizeof(uint32_t))) ||
- (be64toh(*(uint64_t *)&first.addr32) > be64toh(*(uint64_t *)&last.addr32)) ) {
- netdata_log_info("The specified range %s is invalid, the second address is smallest than the first, it will be ignored.",
- ipdup);
- goto cleanipdup;
- }
- } else { // Unique ip
- select = ebpf_ip2nl(first.addr8, ip, AF_INET, ipdup);
- if (select)
- goto cleanipdup;
-
- memcpy(last.addr8, first.addr8, sizeof(first.addr8));
- }
-
- ebpf_network_viewer_ip_list_t *store;
-
- storethisip:
- store = callocz(1, sizeof(ebpf_network_viewer_ip_list_t));
- store->value = ipdup;
- store->hash = simple_hash(ipdup);
- store->ver = (uint8_t)(!is_ipv6)?AF_INET:AF_INET6;
- memcpy(store->first.addr8, first.addr8, sizeof(first.addr8));
- memcpy(store->last.addr8, last.addr8, sizeof(last.addr8));
-
- ebpf_fill_ip_list_unsafe(list, store, "socket");
- return;
-
- cleanipdup:
- freez(ipdup);
-}
-
-/**
- * Parse IP Range
- *
- * Parse the IP ranges given and create Network Viewer IP Structure
- *
- * @param ptr is a pointer with the text to parse.
- */
-void ebpf_parse_ips_unsafe(char *ptr)
-{
- // No value
- if (unlikely(!ptr))
- return;
-
- while (likely(ptr)) {
- // Move forward until next valid character
- while (isspace(*ptr)) ptr++;
-
- // No valid value found
- if (unlikely(!*ptr))
- return;
-
- // Find space that ends the list
- char *end = strchr(ptr, ' ');
- if (end) {
- *end++ = '\0';
- }
-
- int neg = 0;
- if (*ptr == '!') {
- neg++;
- ptr++;
- }
-
- if (isascii(*ptr)) { // Parse port
- ebpf_parse_ip_list_unsafe(
- (!neg) ? (void **)&network_viewer_opt.included_ips : (void **)&network_viewer_opt.excluded_ips, ptr);
- }
-
- ptr = end;
- }
-}
-
-/**
- * Fill Port list
- *
- * @param out a pointer to the link list.
- * @param in the structure that will be linked.
- */
-static inline void fill_port_list(ebpf_network_viewer_port_list_t **out, ebpf_network_viewer_port_list_t *in)
-{
- if (likely(*out)) {
- ebpf_network_viewer_port_list_t *move = *out, *store = *out;
- uint16_t first = ntohs(in->first);
- uint16_t last = ntohs(in->last);
- while (move) {
- uint16_t cmp_first = ntohs(move->first);
- uint16_t cmp_last = ntohs(move->last);
- if (cmp_first <= first && first <= cmp_last &&
- cmp_first <= last && last <= cmp_last ) {
- netdata_log_info("The range/value (%u, %u) is inside the range/value (%u, %u) already inserted, it will be ignored.",
- first, last, cmp_first, cmp_last);
- freez(in->value);
- freez(in);
- return;
- } else if (first <= cmp_first && cmp_first <= last &&
- first <= cmp_last && cmp_last <= last) {
- netdata_log_info("The range (%u, %u) is bigger than previous range (%u, %u) already inserted, the previous will be ignored.",
- first, last, cmp_first, cmp_last);
- freez(move->value);
- move->value = in->value;
- move->first = in->first;
- move->last = in->last;
- freez(in);
- return;
- }
-
- store = move;
- move = move->next;
- }
-
- store->next = in;
- } else {
- *out = in;
- }
-
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("Adding values %s( %u, %u) to %s port list used on network viewer",
- in->value, in->first, in->last,
- (*out == network_viewer_opt.included_port)?"included":"excluded");
-#endif
-}
-
-/**
- * Parse Service List
- *
- * @param out a pointer to store the link list
- * @param service the service used to create the structure that will be linked.
- */
-static void ebpf_parse_service_list(void **out, char *service)
-{
- ebpf_network_viewer_port_list_t **list = (ebpf_network_viewer_port_list_t **)out;
- struct servent *serv = getservbyname((const char *)service, "tcp");
- if (!serv)
- serv = getservbyname((const char *)service, "udp");
-
- if (!serv) {
- netdata_log_info("Cannot resolve the service '%s' with protocols TCP and UDP, it will be ignored", service);
- return;
- }
-
- ebpf_network_viewer_port_list_t *w = callocz(1, sizeof(ebpf_network_viewer_port_list_t));
- w->value = strdupz(service);
- w->hash = simple_hash(service);
-
- w->first = w->last = (uint16_t)serv->s_port;
-
- fill_port_list(list, w);
-}
-
-/**
- * Parse port list
- *
- * Parse an allocated port list with the range given
- *
- * @param out a pointer to store the link list
- * @param range the informed range for the user.
- */
-static void ebpf_parse_port_list(void **out, char *range)
-{
- int first, last;
- ebpf_network_viewer_port_list_t **list = (ebpf_network_viewer_port_list_t **)out;
-
- char *copied = strdupz(range);
- if (*range == '*' && *(range+1) == '\0') {
- first = 1;
- last = 65535;
-
- ebpf_clean_port_structure(list);
- goto fillenvpl;
- }
-
- char *end = range;
- //Move while I cannot find a separator
- while (*end && *end != ':' && *end != '-') end++;
-
- //It has a range
- if (likely(*end)) {
- *end++ = '\0';
- if (*end == '!') {
- netdata_log_info("The exclusion cannot be in the second part of the range, the range %s will be ignored.", copied);
- freez(copied);
- return;
- }
- last = str2i((const char *)end);
- } else {
- last = 0;
- }
-
- first = str2i((const char *)range);
- if (first < NETDATA_MINIMUM_PORT_VALUE || first > NETDATA_MAXIMUM_PORT_VALUE) {
- netdata_log_info("The first port %d of the range \"%s\" is invalid and it will be ignored!", first, copied);
- freez(copied);
- return;
- }
-
- if (!last)
- last = first;
-
- if (last < NETDATA_MINIMUM_PORT_VALUE || last > NETDATA_MAXIMUM_PORT_VALUE) {
- netdata_log_info("The second port %d of the range \"%s\" is invalid and the whole range will be ignored!", last, copied);
- freez(copied);
- return;
- }
-
- if (first > last) {
- netdata_log_info("The specified order %s is wrong, the smallest value is always the first, it will be ignored!", copied);
- freez(copied);
- return;
- }
-
- ebpf_network_viewer_port_list_t *w;
- fillenvpl:
- w = callocz(1, sizeof(ebpf_network_viewer_port_list_t));
- w->value = copied;
- w->hash = simple_hash(copied);
- w->first = (uint16_t)first;
- w->last = (uint16_t)last;
- w->cmp_first = (uint16_t)first;
- w->cmp_last = (uint16_t)last;
-
- fill_port_list(list, w);
-}
-
-/**
- * Parse Port Range
- *
- * Parse the port ranges given and create Network Viewer Port Structure
- *
- * @param ptr is a pointer with the text to parse.
- */
-void ebpf_parse_ports(char *ptr)
-{
- // No value
- if (unlikely(!ptr))
- return;
-
- while (likely(ptr)) {
- // Move forward until next valid character
- while (isspace(*ptr)) ptr++;
-
- // No valid value found
- if (unlikely(!*ptr))
- return;
-
- // Find space that ends the list
- char *end = strchr(ptr, ' ');
- if (end) {
- *end++ = '\0';
- }
-
- int neg = 0;
- if (*ptr == '!') {
- neg++;
- ptr++;
- }
-
- if (isdigit(*ptr)) { // Parse port
- ebpf_parse_port_list(
- (!neg) ? (void **)&network_viewer_opt.included_port : (void **)&network_viewer_opt.excluded_port, ptr);
- } else if (isalpha(*ptr)) { // Parse service
- ebpf_parse_service_list(
- (!neg) ? (void **)&network_viewer_opt.included_port : (void **)&network_viewer_opt.excluded_port, ptr);
- } else if (*ptr == '*') { // All
- ebpf_parse_port_list(
- (!neg) ? (void **)&network_viewer_opt.included_port : (void **)&network_viewer_opt.excluded_port, ptr);
- }
-
- ptr = end;
- }
-}
-
-/*****************************************************************
- *
- * FUNCTIONS TO DEFINE OPTIONS
- *
- *****************************************************************/
-
-/**
- * Define labels used to generate charts
- *
- * @param is structure with information about number of calls made for a function.
- * @param pio structure used to generate charts.
- * @param dim a pointer for the dimensions name
- * @param name a pointer for the tensor with the name of the functions.
- * @param algorithm a vector with the algorithms used to make the charts
- * @param end the number of elements in the previous 4 arguments.
- */
-void ebpf_global_labels(netdata_syscall_stat_t *is, netdata_publish_syscall_t *pio, char **dim,
- char **name, int *algorithm, int end)
-{
- int i;
-
- netdata_syscall_stat_t *prev = NULL;
- netdata_publish_syscall_t *publish_prev = NULL;
- for (i = 0; i < end; i++) {
- if (prev) {
- prev->next = &is[i];
- }
- prev = &is[i];
-
- pio[i].dimension = dim[i];
- pio[i].name = name[i];
- pio[i].algorithm = ebpf_algorithms[algorithm[i]];
- if (publish_prev) {
- publish_prev->next = &pio[i];
- }
- publish_prev = &pio[i];
- }
-}
-
-/**
- * Define thread mode for all ebpf program.
- *
- * @param lmode the mode that will be used for them.
- */
-static inline void ebpf_set_thread_mode(netdata_run_mode_t lmode)
-{
- int i;
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_modules[i].mode = lmode;
- }
-}
-
-/**
- * Enable specific charts selected by user.
- *
- * @param em the structure that will be changed
- * @param disable_cgroup the status about the cgroups charts.
- */
-static inline void ebpf_enable_specific_chart(struct ebpf_module *em, int disable_cgroup)
-{
- em->enabled = NETDATA_THREAD_EBPF_RUNNING;
-
- if (!disable_cgroup) {
- em->cgroup_charts = CONFIG_BOOLEAN_YES;
- }
-
- em->global_charts = CONFIG_BOOLEAN_YES;
-}
-
-/**
- * Disable all Global charts
- *
- * Disable charts
- */
-static inline void disable_all_global_charts()
-{
- int i;
- for (i = 0; ebpf_modules[i].info.thread_name; i++) {
- ebpf_modules[i].enabled = NETDATA_THREAD_EBPF_NOT_RUNNING;
- ebpf_modules[i].global_charts = 0;
- }
-}
-
-/**
- * Enable the specified chart group
- *
- * @param idx the index of ebpf_modules that I am enabling
- */
-static inline void ebpf_enable_chart(int idx, int disable_cgroup)
-{
- int i;
- for (i = 0; ebpf_modules[i].info.thread_name; i++) {
- if (i == idx) {
- ebpf_enable_specific_chart(&ebpf_modules[i], disable_cgroup);
- break;
- }
- }
-}
-
-/**
- * Disable Cgroups
- *
- * Disable charts for apps loading only global charts.
- */
-static inline void ebpf_disable_cgroups()
-{
- int i;
- for (i = 0; ebpf_modules[i].info.thread_name; i++) {
- ebpf_modules[i].cgroup_charts = 0;
- }
-}
-
-/**
- * Update Disabled Plugins
- *
- * This function calls ebpf_update_stats to update statistics for collector.
- *
- * @param em a pointer to `struct ebpf_module`
- */
-void ebpf_update_disabled_plugin_stats(ebpf_module_t *em)
-{
- pthread_mutex_lock(&lock);
- ebpf_update_stats(&plugin_statistics, em);
- pthread_mutex_unlock(&lock);
-}
-
-/**
- * Print help on standard error for user knows how to use the collector.
- */
-void ebpf_print_help()
-{
- const time_t t = time(NULL);
- struct tm ct;
- struct tm *test = localtime_r(&t, &ct);
- int year;
- if (test)
- year = ct.tm_year;
- else
- year = 0;
-
- fprintf(stderr,
- "\n"
- " Netdata ebpf.plugin %s\n"
- " Copyright (C) 2016-%d Costa Tsaousis <costa@tsaousis.gr>\n"
- " Released under GNU General Public License v3 or later.\n"
- " All rights reserved.\n"
- "\n"
- " This eBPF.plugin is a data collector plugin for netdata.\n"
- "\n"
- " This plugin only accepts long options with one or two dashes. The available command line options are:\n"
- "\n"
- " SECONDS Set the data collection frequency.\n"
- "\n"
- " [-]-help Show this help.\n"
- "\n"
- " [-]-version Show software version.\n"
- "\n"
- " [-]-global Disable charts per application and cgroup.\n"
- "\n"
- " [-]-all Enable all chart groups (global, apps, and cgroup), unless -g is also given.\n"
- "\n"
- " [-]-cachestat Enable charts related to process run time.\n"
- "\n"
- " [-]-dcstat Enable charts related to directory cache.\n"
- "\n"
- " [-]-disk Enable charts related to disk monitoring.\n"
- "\n"
- " [-]-filesystem Enable chart related to filesystem run time.\n"
- "\n"
- " [-]-hardirq Enable chart related to hard IRQ latency.\n"
- "\n"
- " [-]-mdflush Enable charts related to multi-device flush.\n"
- "\n"
- " [-]-mount Enable charts related to mount monitoring.\n"
- "\n"
- " [-]-net Enable network viewer charts.\n"
- "\n"
- " [-]-oomkill Enable chart related to OOM kill tracking.\n"
- "\n"
- " [-]-process Enable charts related to process run time.\n"
- "\n"
- " [-]-return Run the collector in return mode.\n"
- "\n"
- " [-]-shm Enable chart related to shared memory tracking.\n"
- "\n"
- " [-]-softirq Enable chart related to soft IRQ latency.\n"
- "\n"
- " [-]-sync Enable chart related to sync run time.\n"
- "\n"
- " [-]-swap Enable chart related to swap run time.\n"
- "\n"
- " [-]-vfs Enable chart related to vfs run time.\n"
- "\n"
- " [-]-legacy Load legacy eBPF programs.\n"
- "\n"
- " [-]-core Use CO-RE when available(Working in progress).\n"
- "\n",
- VERSION,
- (year >= 116) ? year + 1900 : 2020);
-}
-
-/*****************************************************************
- *
- * TRACEPOINT MANAGEMENT FUNCTIONS
- *
- *****************************************************************/
-
-/**
- * Enable a tracepoint.
- *
- * @return 0 on success, -1 on error.
- */
-int ebpf_enable_tracepoint(ebpf_tracepoint_t *tp)
-{
- int test = ebpf_is_tracepoint_enabled(tp->class, tp->event);
-
- // err?
- if (test == -1) {
- return -1;
- }
- // disabled?
- else if (test == 0) {
- // enable it then.
- if (ebpf_enable_tracing_values(tp->class, tp->event)) {
- return -1;
- }
- }
-
- // enabled now or already was.
- tp->enabled = true;
-
- return 0;
-}
-
-/**
- * Disable a tracepoint if it's enabled.
- *
- * @return 0 on success, -1 on error.
- */
-int ebpf_disable_tracepoint(ebpf_tracepoint_t *tp)
-{
- int test = ebpf_is_tracepoint_enabled(tp->class, tp->event);
-
- // err?
- if (test == -1) {
- return -1;
- }
- // enabled?
- else if (test == 1) {
- // disable it then.
- if (ebpf_disable_tracing_values(tp->class, tp->event)) {
- return -1;
- }
- }
-
- // disable now or already was.
- tp->enabled = false;
-
- return 0;
-}
-
-/**
- * Enable multiple tracepoints on a list of tracepoints which end when the
- * class is NULL.
- *
- * @return the number of successful enables.
- */
-uint32_t ebpf_enable_tracepoints(ebpf_tracepoint_t *tps)
-{
- uint32_t cnt = 0;
- for (int i = 0; tps[i].class != NULL; i++) {
- if (ebpf_enable_tracepoint(&tps[i]) == -1) {
- netdata_log_error("Failed to enable tracepoint %s:%s", tps[i].class, tps[i].event);
- }
- else {
- cnt += 1;
- }
- }
- return cnt;
-}
-
-/*****************************************************************
- *
- * AUXILIARY FUNCTIONS USED DURING INITIALIZATION
- *
- *****************************************************************/
-
-/**
- * Is ip inside the range
- *
- * Check if the ip is inside a IP range
- *
- * @param rfirst the first ip address of the range
- * @param rlast the last ip address of the range
- * @param cmpfirst the first ip to compare
- * @param cmplast the last ip to compare
- * @param family the IP family
- *
- * @return It returns 1 if the IP is inside the range and 0 otherwise
- */
-static int ebpf_is_ip_inside_range(union netdata_ip_t *rfirst, union netdata_ip_t *rlast,
- union netdata_ip_t *cmpfirst, union netdata_ip_t *cmplast, int family)
-{
- if (family == AF_INET) {
- if ((rfirst->addr32[0] <= cmpfirst->addr32[0]) && (rlast->addr32[0] >= cmplast->addr32[0]))
- return 1;
- } else {
- if (memcmp(rfirst->addr8, cmpfirst->addr8, sizeof(union netdata_ip_t)) <= 0 &&
- memcmp(rlast->addr8, cmplast->addr8, sizeof(union netdata_ip_t)) >= 0) {
- return 1;
- }
-
- }
- return 0;
-}
-
-/**
- * Fill IP list
- *
- * @param out a pointer to the link list.
- * @param in the structure that will be linked.
- * @param table the modified table.
- */
-void ebpf_fill_ip_list_unsafe(ebpf_network_viewer_ip_list_t **out, ebpf_network_viewer_ip_list_t *in,
- char *table __maybe_unused)
-{
- if (in->ver == AF_INET) { // It is simpler to compare using host order
- in->first.addr32[0] = ntohl(in->first.addr32[0]);
- in->last.addr32[0] = ntohl(in->last.addr32[0]);
- }
- if (likely(*out)) {
- ebpf_network_viewer_ip_list_t *move = *out, *store = *out;
- while (move) {
- if (in->ver == move->ver &&
- ebpf_is_ip_inside_range(&move->first, &move->last, &in->first, &in->last, in->ver)) {
-#ifdef NETDATA_DEV_MODE
- netdata_log_info("The range/value (%s) is inside the range/value (%s) already inserted, it will be ignored.",
- in->value, move->value);
-#endif
- freez(in->value);
- freez(in);
- return;
- }
- store = move;
- move = move->next;
- }
-
- store->next = in;
- } else {
- *out = in;
- }
-
-#ifdef NETDATA_DEV_MODE
- char first[256], last[512];
- if (in->ver == AF_INET) {
- netdata_log_info("Adding values %s: (%u - %u) to %s IP list \"%s\" used on network viewer",
- in->value, in->first.addr32[0], in->last.addr32[0],
- (*out == network_viewer_opt.included_ips)?"included":"excluded",
- table);
- } else {
- if (inet_ntop(AF_INET6, in->first.addr8, first, INET6_ADDRSTRLEN) &&
- inet_ntop(AF_INET6, in->last.addr8, last, INET6_ADDRSTRLEN))
- netdata_log_info("Adding values %s - %s to %s IP list \"%s\" used on network viewer",
- first, last,
- (*out == network_viewer_opt.included_ips)?"included":"excluded",
- table);
- }
-#endif
-}
-
-/**
- * Link hostname
- *
- * @param out is the output link list
- * @param in the hostname to add to list.
- */
-static void ebpf_link_hostname(ebpf_network_viewer_hostname_list_t **out, ebpf_network_viewer_hostname_list_t *in)
-{
- if (likely(*out)) {
- ebpf_network_viewer_hostname_list_t *move = *out;
- for (; move->next ; move = move->next ) {
- if (move->hash == in->hash && !strcmp(move->value, in->value)) {
- netdata_log_info("The hostname %s was already inserted, it will be ignored.", in->value);
- freez(in->value);
- simple_pattern_free(in->value_pattern);
- freez(in);
- return;
- }
- }
-
- move->next = in;
- } else {
- *out = in;
- }
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("Adding value %s to %s hostname list used on network viewer",
- in->value,
- (*out == network_viewer_opt.included_hostnames)?"included":"excluded");
-#endif
-}
-
-/**
- * Link Hostnames
- *
- * Parse the list of hostnames to create the link list.
- * This is not associated with the IP, because simple patterns like *example* cannot be resolved to IP.
- *
- * @param out is the output link list
- * @param parse is a pointer with the text to parser.
- */
-static void ebpf_link_hostnames(char *parse)
-{
- // No value
- if (unlikely(!parse))
- return;
-
- while (likely(parse)) {
- // Find the first valid value
- while (isspace(*parse)) parse++;
-
- // No valid value found
- if (unlikely(!*parse))
- return;
-
- // Find space that ends the list
- char *end = strchr(parse, ' ');
- if (end) {
- *end++ = '\0';
- }
-
- int neg = 0;
- if (*parse == '!') {
- neg++;
- parse++;
- }
-
- ebpf_network_viewer_hostname_list_t *hostname = callocz(1 , sizeof(ebpf_network_viewer_hostname_list_t));
- hostname->value = strdupz(parse);
- hostname->hash = simple_hash(parse);
- hostname->value_pattern = simple_pattern_create(parse, NULL, SIMPLE_PATTERN_EXACT, true);
-
- ebpf_link_hostname((!neg) ? &network_viewer_opt.included_hostnames :
- &network_viewer_opt.excluded_hostnames,
- hostname);
-
- parse = end;
- }
-}
-
-/**
- * Parse network viewer section
- *
- * @param cfg the configuration structure
- */
-void parse_network_viewer_section(struct config *cfg)
-{
- network_viewer_opt.hostname_resolution_enabled = appconfig_get_boolean(cfg,
- EBPF_NETWORK_VIEWER_SECTION,
- EBPF_CONFIG_RESOLVE_HOSTNAME,
- CONFIG_BOOLEAN_NO);
-
- network_viewer_opt.service_resolution_enabled = appconfig_get_boolean(cfg,
- EBPF_NETWORK_VIEWER_SECTION,
- EBPF_CONFIG_RESOLVE_SERVICE,
- CONFIG_BOOLEAN_YES);
-
- char *value = appconfig_get(cfg, EBPF_NETWORK_VIEWER_SECTION, EBPF_CONFIG_PORTS, NULL);
- ebpf_parse_ports(value);
-
- if (network_viewer_opt.hostname_resolution_enabled) {
- value = appconfig_get(cfg, EBPF_NETWORK_VIEWER_SECTION, EBPF_CONFIG_HOSTNAMES, NULL);
- ebpf_link_hostnames(value);
- } else {
- netdata_log_info("Name resolution is disabled, collector will not parse \"hostnames\" list.");
- }
-
- value = appconfig_get(cfg,
- EBPF_NETWORK_VIEWER_SECTION,
- "ips",
- NULL);
- //"ips", "!127.0.0.1/8 10.0.0.0/8 172.16.0.0/12 192.168.0.0/16 fc00::/7 !::1/128");
- ebpf_parse_ips_unsafe(value);
-}
-
-/**
- * Read Local Ports
- *
- * Parse /proc/net/{tcp,udp} and get the ports Linux is listening.
- *
- * @param filename the proc file to parse.
- * @param proto is the magic number associated to the protocol file we are reading.
- */
-static void read_local_ports(char *filename, uint8_t proto)
-{
- procfile *ff = procfile_open(filename, " \t:", PROCFILE_FLAG_DEFAULT);
- if (!ff)
- return;
-
- ff = procfile_readall(ff);
- if (!ff)
- return;
-
- size_t lines = procfile_lines(ff), l;
- netdata_passive_connection_t values = {.counter = 0, .tgid = 0, .pid = 0};
- for(l = 0; l < lines ;l++) {
- size_t words = procfile_linewords(ff, l);
- // This is header or end of file
- if (unlikely(words < 14))
- continue;
-
- // https://elixir.bootlin.com/linux/v5.7.8/source/include/net/tcp_states.h
- // 0A = TCP_LISTEN
- if (strcmp("0A", procfile_lineword(ff, l, 5)))
- continue;
-
- // Read local port
- uint16_t port = (uint16_t)strtol(procfile_lineword(ff, l, 2), NULL, 16);
- update_listen_table(htons(port), proto, &values);
- }
-
- procfile_close(ff);
-}
-
-/**
- * Read Local addresseses
- *
- * Read the local address from the interfaces.
- */
-void ebpf_read_local_addresses_unsafe()
-{
- struct ifaddrs *ifaddr, *ifa;
- if (getifaddrs(&ifaddr) == -1) {
- netdata_log_error("Cannot get the local IP addresses, it is no possible to do separation between inbound and outbound connections");
- return;
- }
-
- char *notext = { "No text representation" };
- for (ifa = ifaddr; ifa != NULL; ifa = ifa->ifa_next) {
- if (ifa->ifa_addr == NULL)
- continue;
-
- if ((ifa->ifa_addr->sa_family != AF_INET) && (ifa->ifa_addr->sa_family != AF_INET6))
- continue;
-
- ebpf_network_viewer_ip_list_t *w = callocz(1, sizeof(ebpf_network_viewer_ip_list_t));
-
- int family = ifa->ifa_addr->sa_family;
- w->ver = (uint8_t) family;
- char text[INET6_ADDRSTRLEN];
- if (family == AF_INET) {
- struct sockaddr_in *in = (struct sockaddr_in*) ifa->ifa_addr;
-
- w->first.addr32[0] = in->sin_addr.s_addr;
- w->last.addr32[0] = in->sin_addr.s_addr;
-
- if (inet_ntop(AF_INET, w->first.addr8, text, INET_ADDRSTRLEN)) {
- w->value = strdupz(text);
- w->hash = simple_hash(text);
- } else {
- w->value = strdupz(notext);
- w->hash = simple_hash(notext);
- }
- } else {
- struct sockaddr_in6 *in6 = (struct sockaddr_in6*) ifa->ifa_addr;
-
- memcpy(w->first.addr8, (void *)&in6->sin6_addr, sizeof(struct in6_addr));
- memcpy(w->last.addr8, (void *)&in6->sin6_addr, sizeof(struct in6_addr));
-
- if (inet_ntop(AF_INET6, w->first.addr8, text, INET_ADDRSTRLEN)) {
- w->value = strdupz(text);
- w->hash = simple_hash(text);
- } else {
- w->value = strdupz(notext);
- w->hash = simple_hash(notext);
- }
- }
-
- ebpf_fill_ip_list_unsafe(
- (family == AF_INET) ? &network_viewer_opt.ipv4_local_ip : &network_viewer_opt.ipv6_local_ip, w, "selector");
- }
-
- freeifaddrs(ifaddr);
-}
-
-/**
- * Start Pthread Variable
- *
- * This function starts all pthread variables.
- */
-void ebpf_start_pthread_variables()
-{
- pthread_mutex_init(&lock, NULL);
- pthread_mutex_init(&ebpf_exit_cleanup, NULL);
- pthread_mutex_init(&collect_data_mutex, NULL);
- pthread_mutex_init(&mutex_cgroup_shm, NULL);
- rw_spinlock_init(&ebpf_judy_pid.index.rw_spinlock);
-}
-
-/**
- * Allocate the vectors used for all threads.
- */
-static void ebpf_allocate_common_vectors()
-{
- ebpf_judy_pid.pid_table = ebpf_allocate_pid_aral(NETDATA_EBPF_PID_SOCKET_ARAL_TABLE_NAME,
- sizeof(netdata_ebpf_judy_pid_stats_t));
- ebpf_all_pids = callocz((size_t)pid_max, sizeof(struct ebpf_pid_stat *));
- ebpf_aral_init();
-}
-
-/**
- * Define how to load the ebpf programs
- *
- * @param ptr the option given by users
- */
-static inline void ebpf_how_to_load(char *ptr)
-{
- if (!strcasecmp(ptr, EBPF_CFG_LOAD_MODE_RETURN))
- ebpf_set_thread_mode(MODE_RETURN);
- else if (!strcasecmp(ptr, EBPF_CFG_LOAD_MODE_DEFAULT))
- ebpf_set_thread_mode(MODE_ENTRY);
- else
- netdata_log_error("the option %s for \"ebpf load mode\" is not a valid option.", ptr);
-}
-
-/**
- * Define whether we should have charts for apps
- *
- * @param lmode the mode that will be used for them.
- */
-static inline void ebpf_set_apps_mode(netdata_apps_integration_flags_t value)
-{
- int i;
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_modules[i].apps_charts = value;
- }
-}
-
-
-/**
- * Update interval
- *
- * Update default interval with value from user
- *
- * @param update_every value to overwrite the update frequency set by the server.
- */
-static void ebpf_update_interval(int update_every)
-{
- int i;
- int value = (int) appconfig_get_number(&collector_config, EBPF_GLOBAL_SECTION, EBPF_CFG_UPDATE_EVERY,
- update_every);
- for (i = 0; ebpf_modules[i].info.thread_name; i++) {
- ebpf_modules[i].update_every = value;
- }
-}
-
-/**
- * Update PID table size
- *
- * Update default size with value from user
- */
-static void ebpf_update_table_size()
-{
- int i;
- uint32_t value = (uint32_t) appconfig_get_number(&collector_config, EBPF_GLOBAL_SECTION,
- EBPF_CFG_PID_SIZE, ND_EBPF_DEFAULT_PID_SIZE);
- for (i = 0; ebpf_modules[i].info.thread_name; i++) {
- ebpf_modules[i].pid_map_size = value;
- }
-}
-
-/**
- * Update lifetime
- *
- * Update the period of time that specific thread will run
- */
-static void ebpf_update_lifetime()
-{
- int i;
- uint32_t value = (uint32_t) appconfig_get_number(&collector_config, EBPF_GLOBAL_SECTION,
- EBPF_CFG_LIFETIME, EBPF_DEFAULT_LIFETIME);
- for (i = 0; ebpf_modules[i].info.thread_name; i++) {
- ebpf_modules[i].lifetime = value;
- }
-}
-
-/**
- * Set Load mode
- *
- * @param origin specify the configuration file loaded
- */
-static inline void ebpf_set_load_mode(netdata_ebpf_load_mode_t load, netdata_ebpf_load_mode_t origin)
-{
- int i;
- for (i = 0; ebpf_modules[i].info.thread_name; i++) {
- ebpf_modules[i].load &= ~NETDATA_EBPF_LOAD_METHODS;
- ebpf_modules[i].load |= load | origin ;
- }
-}
-
-/**
- * Update mode
- *
- * @param str value read from configuration file.
- * @param origin specify the configuration file loaded
- */
-static inline void epbf_update_load_mode(char *str, netdata_ebpf_load_mode_t origin)
-{
- netdata_ebpf_load_mode_t load = epbf_convert_string_to_load_mode(str);
-
- ebpf_set_load_mode(load, origin);
-}
-
-/**
- * Update Map per core
- *
- * Define the map type used with some hash tables.
- */
-static void ebpf_update_map_per_core()
-{
- int i;
- int value = appconfig_get_boolean(&collector_config, EBPF_GLOBAL_SECTION,
- EBPF_CFG_MAPS_PER_CORE, CONFIG_BOOLEAN_YES);
- for (i = 0; ebpf_modules[i].info.thread_name; i++) {
- ebpf_modules[i].maps_per_core = value;
- }
-}
-
-/**
- * Read collector values
- *
- * @param disable_cgroups variable to store information related to cgroups.
- * @param update_every value to overwrite the update frequency set by the server.
- * @param origin specify the configuration file loaded
- */
-static void read_collector_values(int *disable_cgroups,
- int update_every, netdata_ebpf_load_mode_t origin)
-{
- // Read global section
- char *value;
- if (appconfig_exists(&collector_config, EBPF_GLOBAL_SECTION, "load")) // Backward compatibility
- value = appconfig_get(&collector_config, EBPF_GLOBAL_SECTION, "load",
- EBPF_CFG_LOAD_MODE_DEFAULT);
- else
- value = appconfig_get(&collector_config, EBPF_GLOBAL_SECTION, EBPF_CFG_LOAD_MODE,
- EBPF_CFG_LOAD_MODE_DEFAULT);
-
- ebpf_how_to_load(value);
-
- btf_path = appconfig_get(&collector_config, EBPF_GLOBAL_SECTION, EBPF_CFG_PROGRAM_PATH,
- EBPF_DEFAULT_BTF_PATH);
-
-#ifdef LIBBPF_MAJOR_VERSION
- default_btf = ebpf_load_btf_file(btf_path, EBPF_DEFAULT_BTF_FILE);
-#endif
-
- value = appconfig_get(&collector_config, EBPF_GLOBAL_SECTION, EBPF_CFG_TYPE_FORMAT, EBPF_CFG_DEFAULT_PROGRAM);
-
- epbf_update_load_mode(value, origin);
-
- ebpf_update_interval(update_every);
-
- ebpf_update_table_size();
-
- ebpf_update_lifetime();
-
- // This is kept to keep compatibility
- uint32_t enabled = appconfig_get_boolean(&collector_config, EBPF_GLOBAL_SECTION, "disable apps",
- CONFIG_BOOLEAN_NO);
- if (!enabled) {
- // Apps is a positive sentence, so we need to invert the values to disable apps.
- enabled = appconfig_get_boolean(&collector_config, EBPF_GLOBAL_SECTION, EBPF_CFG_APPLICATION,
- CONFIG_BOOLEAN_YES);
- enabled = (enabled == CONFIG_BOOLEAN_NO)?CONFIG_BOOLEAN_YES:CONFIG_BOOLEAN_NO;
- }
-
- ebpf_set_apps_mode(!enabled);
-
- // Cgroup is a positive sentence, so we need to invert the values to disable apps.
- // We are using the same pattern for cgroup and apps
- enabled = appconfig_get_boolean(&collector_config, EBPF_GLOBAL_SECTION, EBPF_CFG_CGROUP, CONFIG_BOOLEAN_NO);
- *disable_cgroups = (enabled == CONFIG_BOOLEAN_NO)?CONFIG_BOOLEAN_YES:CONFIG_BOOLEAN_NO;
-
- ebpf_update_map_per_core();
-
- // Read ebpf programs section
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION,
- ebpf_modules[EBPF_MODULE_PROCESS_IDX].info.config_name, CONFIG_BOOLEAN_YES);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_PROCESS_IDX, *disable_cgroups);
- }
-
- // This is kept to keep compatibility
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "network viewer",
- CONFIG_BOOLEAN_NO);
- if (!enabled)
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION,
- ebpf_modules[EBPF_MODULE_SOCKET_IDX].info.config_name,
- CONFIG_BOOLEAN_NO);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_SOCKET_IDX, *disable_cgroups);
- }
-
- // This is kept to keep compatibility
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "network connection monitoring",
- CONFIG_BOOLEAN_YES);
- if (!enabled)
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "network connections",
- CONFIG_BOOLEAN_YES);
-
- network_viewer_opt.enabled = enabled;
- if (enabled) {
- if (!ebpf_modules[EBPF_MODULE_SOCKET_IDX].enabled)
- ebpf_enable_chart(EBPF_MODULE_SOCKET_IDX, *disable_cgroups);
-
- // Read network viewer section if network viewer is enabled
- // This is kept here to keep backward compatibility
- parse_network_viewer_section(&collector_config);
- ebpf_parse_service_name_section(&collector_config);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "cachestat",
- CONFIG_BOOLEAN_NO);
-
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_CACHESTAT_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "sync",
- CONFIG_BOOLEAN_YES);
-
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_SYNC_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "dcstat",
- CONFIG_BOOLEAN_NO);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_DCSTAT_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "swap",
- CONFIG_BOOLEAN_NO);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_SWAP_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "vfs",
- CONFIG_BOOLEAN_NO);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_VFS_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "filesystem",
- CONFIG_BOOLEAN_NO);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_FILESYSTEM_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "disk",
- CONFIG_BOOLEAN_NO);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_DISK_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "mount",
- CONFIG_BOOLEAN_YES);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_MOUNT_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "fd",
- CONFIG_BOOLEAN_YES);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_FD_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "hardirq",
- CONFIG_BOOLEAN_YES);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_HARDIRQ_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "softirq",
- CONFIG_BOOLEAN_YES);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_SOFTIRQ_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "oomkill",
- CONFIG_BOOLEAN_YES);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_OOMKILL_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "shm",
- CONFIG_BOOLEAN_YES);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_SHM_IDX, *disable_cgroups);
- }
-
- enabled = appconfig_get_boolean(&collector_config, EBPF_PROGRAMS_SECTION, "mdflush",
- CONFIG_BOOLEAN_NO);
- if (enabled) {
- ebpf_enable_chart(EBPF_MODULE_MDFLUSH_IDX, *disable_cgroups);
- }
-}
-
-/**
- * Load collector config
- *
- * @param path the path where the file ebpf.conf is stored.
- * @param disable_cgroups variable to store the information about cgroups plugin status.
- * @param update_every value to overwrite the update frequency set by the server.
- *
- * @return 0 on success and -1 otherwise.
- */
-static int ebpf_load_collector_config(char *path, int *disable_cgroups, int update_every)
-{
- char lpath[4096];
- netdata_ebpf_load_mode_t origin;
-
- snprintf(lpath, 4095, "%s/%s", path, NETDATA_EBPF_CONFIG_FILE);
- if (!appconfig_load(&collector_config, lpath, 0, NULL)) {
- snprintf(lpath, 4095, "%s/%s", path, NETDATA_EBPF_OLD_CONFIG_FILE);
- if (!appconfig_load(&collector_config, lpath, 0, NULL)) {
- return -1;
- }
- origin = EBPF_LOADED_FROM_STOCK;
- } else
- origin = EBPF_LOADED_FROM_USER;
-
- read_collector_values(disable_cgroups, update_every, origin);
-
- return 0;
-}
-
-/**
- * Set global variables reading environment variables
- */
-void set_global_variables()
-{
- // Get environment variables
- ebpf_plugin_dir = getenv("NETDATA_PLUGINS_DIR");
- if (!ebpf_plugin_dir)
- ebpf_plugin_dir = PLUGINS_DIR;
-
- ebpf_user_config_dir = getenv("NETDATA_USER_CONFIG_DIR");
- if (!ebpf_user_config_dir)
- ebpf_user_config_dir = CONFIG_DIR;
-
- ebpf_stock_config_dir = getenv("NETDATA_STOCK_CONFIG_DIR");
- if (!ebpf_stock_config_dir)
- ebpf_stock_config_dir = LIBCONFIG_DIR;
-
- ebpf_configured_log_dir = getenv("NETDATA_LOG_DIR");
- if (!ebpf_configured_log_dir)
- ebpf_configured_log_dir = LOG_DIR;
-
- ebpf_nprocs = (int)sysconf(_SC_NPROCESSORS_ONLN);
- if (ebpf_nprocs < 0) {
- ebpf_nprocs = NETDATA_MAX_PROCESSOR;
- netdata_log_error("Cannot identify number of process, using default value %d", ebpf_nprocs);
- }
-
- isrh = get_redhat_release();
- pid_max = get_system_pid_max();
- running_on_kernel = ebpf_get_kernel_version();
-}
-
-/**
- * Load collector config
- */
-static inline void ebpf_load_thread_config()
-{
- int i;
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_update_module(&ebpf_modules[i], default_btf, running_on_kernel, isrh);
- }
-}
-
-/**
- * Check Conditions
- *
- * This function checks kernel that plugin is running and permissions.
- *
- * @return It returns 0 on success and -1 otherwise
- */
-int ebpf_check_conditions()
-{
- if (!has_condition_to_run(running_on_kernel)) {
- netdata_log_error("The current collector cannot run on this kernel.");
- return -1;
- }
-
- if (!am_i_running_as_root()) {
- netdata_log_error(
- "ebpf.plugin should either run as root (now running with uid %u, euid %u) or have special capabilities..",
- (unsigned int)getuid(), (unsigned int)geteuid());
- return -1;
- }
-
- return 0;
-}
-
-/**
- * Adjust memory
- *
- * Adjust memory values to load eBPF programs.
- *
- * @return It returns 0 on success and -1 otherwise
- */
-int ebpf_adjust_memory_limit()
-{
- struct rlimit r = { RLIM_INFINITY, RLIM_INFINITY };
- if (setrlimit(RLIMIT_MEMLOCK, &r)) {
- netdata_log_error("Setrlimit(RLIMIT_MEMLOCK)");
- return -1;
- }
-
- return 0;
-}
-
-/**
- * Parse arguments given from user.
- *
- * @param argc the number of arguments
- * @param argv the pointer to the arguments
- */
-static void ebpf_parse_args(int argc, char **argv)
-{
- int disable_cgroups = 1;
- int freq = 0;
- int option_index = 0;
- uint64_t select_threads = 0;
- static struct option long_options[] = {
- {"process", no_argument, 0, 0 },
- {"net", no_argument, 0, 0 },
- {"cachestat", no_argument, 0, 0 },
- {"sync", no_argument, 0, 0 },
- {"dcstat", no_argument, 0, 0 },
- {"swap", no_argument, 0, 0 },
- {"vfs", no_argument, 0, 0 },
- {"filesystem", no_argument, 0, 0 },
- {"disk", no_argument, 0, 0 },
- {"mount", no_argument, 0, 0 },
- {"filedescriptor", no_argument, 0, 0 },
- {"hardirq", no_argument, 0, 0 },
- {"softirq", no_argument, 0, 0 },
- {"oomkill", no_argument, 0, 0 },
- {"shm", no_argument, 0, 0 },
- {"mdflush", no_argument, 0, 0 },
- /* INSERT NEW THREADS BEFORE THIS COMMENT TO KEEP COMPATIBILITY WITH enum ebpf_module_indexes */
- {"all", no_argument, 0, 0 },
- {"version", no_argument, 0, 0 },
- {"help", no_argument, 0, 0 },
- {"global", no_argument, 0, 0 },
- {"return", no_argument, 0, 0 },
- {"legacy", no_argument, 0, 0 },
- {"core", no_argument, 0, 0 },
- {"unittest", no_argument, 0, 0 },
- {0, 0, 0, 0}
- };
-
- memset(&network_viewer_opt, 0, sizeof(network_viewer_opt));
- rw_spinlock_init(&network_viewer_opt.rw_spinlock);
-
- if (argc > 1) {
- int n = (int)str2l(argv[1]);
- if (n > 0) {
- freq = n;
- }
- }
-
- if (!freq)
- freq = EBPF_DEFAULT_UPDATE_EVERY;
-
- //rw_spinlock_write_lock(&network_viewer_opt.rw_spinlock);
- if (ebpf_load_collector_config(ebpf_user_config_dir, &disable_cgroups, freq)) {
- netdata_log_info(
- "Does not have a configuration file inside `%s/ebpf.d.conf. It will try to load stock file.",
- ebpf_user_config_dir);
- if (ebpf_load_collector_config(ebpf_stock_config_dir, &disable_cgroups, freq)) {
- netdata_log_info("Does not have a stock file. It is starting with default options.");
- }
- }
-
- ebpf_load_thread_config();
- //rw_spinlock_write_unlock(&network_viewer_opt.rw_spinlock);
-
- while (1) {
- int c = getopt_long_only(argc, argv, "", long_options, &option_index);
- if (c == -1)
- break;
-
- switch (option_index) {
- case EBPF_MODULE_PROCESS_IDX: {
- select_threads |= 1<<EBPF_MODULE_PROCESS_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"PROCESS\" charts, because it was started with the option \"[-]-process\".");
-#endif
- break;
- }
- case EBPF_MODULE_SOCKET_IDX: {
- select_threads |= 1<<EBPF_MODULE_SOCKET_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"NET\" charts, because it was started with the option \"[-]-net\".");
-#endif
- break;
- }
- case EBPF_MODULE_CACHESTAT_IDX: {
- select_threads |= 1<<EBPF_MODULE_CACHESTAT_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"CACHESTAT\" charts, because it was started with the option \"[-]-cachestat\".");
-#endif
- break;
- }
- case EBPF_MODULE_SYNC_IDX: {
- select_threads |= 1<<EBPF_MODULE_SYNC_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"SYNC\" chart, because it was started with the option \"[-]-sync\".");
-#endif
- break;
- }
- case EBPF_MODULE_DCSTAT_IDX: {
- select_threads |= 1<<EBPF_MODULE_DCSTAT_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"DCSTAT\" charts, because it was started with the option \"[-]-dcstat\".");
-#endif
- break;
- }
- case EBPF_MODULE_SWAP_IDX: {
- select_threads |= 1<<EBPF_MODULE_SWAP_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"SWAP\" chart, because it was started with the option \"[-]-swap\".");
-#endif
- break;
- }
- case EBPF_MODULE_VFS_IDX: {
- select_threads |= 1<<EBPF_MODULE_VFS_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"VFS\" chart, because it was started with the option \"[-]-vfs\".");
-#endif
- break;
- }
- case EBPF_MODULE_FILESYSTEM_IDX: {
- select_threads |= 1<<EBPF_MODULE_FILESYSTEM_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"FILESYSTEM\" chart, because it was started with the option \"[-]-filesystem\".");
-#endif
- break;
- }
- case EBPF_MODULE_DISK_IDX: {
- select_threads |= 1<<EBPF_MODULE_DISK_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"DISK\" chart, because it was started with the option \"[-]-disk\".");
-#endif
- break;
- }
- case EBPF_MODULE_MOUNT_IDX: {
- select_threads |= 1<<EBPF_MODULE_MOUNT_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"MOUNT\" chart, because it was started with the option \"[-]-mount\".");
-#endif
- break;
- }
- case EBPF_MODULE_FD_IDX: {
- select_threads |= 1<<EBPF_MODULE_FD_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"FILEDESCRIPTOR\" chart, because it was started with the option \"[-]-filedescriptor\".");
-#endif
- break;
- }
- case EBPF_MODULE_HARDIRQ_IDX: {
- select_threads |= 1<<EBPF_MODULE_HARDIRQ_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"HARDIRQ\" chart, because it was started with the option \"[-]-hardirq\".");
-#endif
- break;
- }
- case EBPF_MODULE_SOFTIRQ_IDX: {
- select_threads |= 1<<EBPF_MODULE_SOFTIRQ_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"SOFTIRQ\" chart, because it was started with the option \"[-]-softirq\".");
-#endif
- break;
- }
- case EBPF_MODULE_OOMKILL_IDX: {
- select_threads |= 1<<EBPF_MODULE_OOMKILL_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"OOMKILL\" chart, because it was started with the option \"[-]-oomkill\".");
-#endif
- break;
- }
- case EBPF_MODULE_SHM_IDX: {
- select_threads |= 1<<EBPF_MODULE_SHM_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"SHM\" chart, because it was started with the option \"[-]-shm\".");
-#endif
- break;
- }
- case EBPF_MODULE_MDFLUSH_IDX: {
- select_threads |= 1<<EBPF_MODULE_MDFLUSH_IDX;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF enabling \"MDFLUSH\" chart, because it was started with the option \"[-]-mdflush\".");
-#endif
- break;
- }
- case EBPF_OPTION_ALL_CHARTS: {
- ebpf_set_apps_mode(NETDATA_EBPF_APPS_FLAG_YES);
- disable_cgroups = 0;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF running with all chart groups, because it was started with the option \"[-]-all\".");
-#endif
- break;
- }
- case EBPF_OPTION_VERSION: {
- printf("ebpf.plugin %s\n", VERSION);
- exit(0);
- }
- case EBPF_OPTION_HELP: {
- ebpf_print_help();
- exit(0);
- }
- case EBPF_OPTION_GLOBAL_CHART: {
- disable_cgroups = 1;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF running with global chart group, because it was started with the option \"[-]-global\".");
-#endif
- break;
- }
- case EBPF_OPTION_RETURN_MODE: {
- ebpf_set_thread_mode(MODE_RETURN);
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF running in \"RETURN\" mode, because it was started with the option \"[-]-return\".");
-#endif
- break;
- }
- case EBPF_OPTION_LEGACY: {
- ebpf_set_load_mode(EBPF_LOAD_LEGACY, EBPF_LOADED_FROM_USER);
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF running with \"LEGACY\" code, because it was started with the option \"[-]-legacy\".");
-#endif
- break;
- }
- case EBPF_OPTION_CORE: {
- ebpf_set_load_mode(EBPF_LOAD_CORE, EBPF_LOADED_FROM_USER);
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("EBPF running with \"CO-RE\" code, because it was started with the option \"[-]-core\".");
-#endif
- break;
- }
- case EBPF_OPTION_UNITTEST: {
- // if we cannot run until the end, we will cancel the unittest
- int exit_code = ECANCELED;
- if (ebpf_check_conditions())
- goto unittest;
-
- if (ebpf_adjust_memory_limit())
- goto unittest;
-
- // Load binary in entry mode
- ebpf_ut_initialize_structure(MODE_ENTRY);
- if (ebpf_ut_load_real_binary())
- goto unittest;
-
- ebpf_ut_cleanup_memory();
-
- // Do not load a binary in entry mode
- ebpf_ut_initialize_structure(MODE_ENTRY);
- if (ebpf_ut_load_fake_binary())
- goto unittest;
-
- ebpf_ut_cleanup_memory();
-
- exit_code = 0;
-unittest:
- exit(exit_code);
- }
- default: {
- break;
- }
- }
- }
-
- if (disable_cgroups) {
- ebpf_disable_cgroups();
- }
-
- if (select_threads) {
- disable_all_global_charts();
- uint64_t idx;
- for (idx = 0; idx < EBPF_OPTION_ALL_CHARTS; idx++) {
- if (select_threads & 1<<idx)
- ebpf_enable_specific_chart(&ebpf_modules[idx], disable_cgroups);
- }
- }
-
- // Load apps_groups.conf
- if (ebpf_read_apps_groups_conf(
- &apps_groups_default_target, &apps_groups_root_target, ebpf_user_config_dir, "groups")) {
- netdata_log_info("Cannot read process groups configuration file '%s/apps_groups.conf'. Will try '%s/apps_groups.conf'",
- ebpf_user_config_dir, ebpf_stock_config_dir);
- if (ebpf_read_apps_groups_conf(
- &apps_groups_default_target, &apps_groups_root_target, ebpf_stock_config_dir, "groups")) {
- netdata_log_error("Cannot read process groups '%s/apps_groups.conf'. There are no internal defaults. Failing.",
- ebpf_stock_config_dir);
- ebpf_exit();
- }
- } else
- netdata_log_info("Loaded config file '%s/apps_groups.conf'", ebpf_user_config_dir);
-}
-
-/*****************************************************************
- *
- * Collector charts
- *
- *****************************************************************/
-
-static char *load_event_stat[NETDATA_EBPF_LOAD_STAT_END] = {"legacy", "co-re"};
-static char *memlock_stat = {"memory_locked"};
-static char *hash_table_stat = {"hash_table"};
-static char *hash_table_core[NETDATA_EBPF_LOAD_STAT_END] = {"per_core", "unique"};
-
-/**
- * Send Hash Table PID data
- *
- * Send all information associated with a specific pid table.
- *
- * @param chart chart id
- * @param idx index position in hash_table_stats
- */
-static inline void ebpf_send_hash_table_pid_data(char *chart, uint32_t idx)
-{
- int i;
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, chart, "");
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_module_t *wem = &ebpf_modules[i];
- if (wem->functions.apps_routine)
- write_chart_dimension((char *)wem->info.thread_name,
- (wem->enabled < NETDATA_THREAD_EBPF_STOPPING) ?
- wem->hash_table_stats[idx]:
- 0);
- }
- ebpf_write_end_chart();
-}
-
-/**
- * Send Global Hash Table data
- *
- * Send all information associated with a specific pid table.
- *
- */
-static inline void ebpf_send_global_hash_table_data()
-{
- int i;
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, NETDATA_EBPF_HASH_TABLES_GLOBAL_ELEMENTS, "");
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_module_t *wem = &ebpf_modules[i];
- write_chart_dimension((char *)wem->info.thread_name,
- (wem->enabled < NETDATA_THREAD_EBPF_STOPPING) ? NETDATA_CONTROLLER_END: 0);
- }
- ebpf_write_end_chart();
-}
-
-/**
- * Send Statistic Data
- *
- * Send statistic information to netdata.
- */
-void ebpf_send_statistic_data()
-{
- if (!publish_internal_metrics)
- return;
-
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, NETDATA_EBPF_THREADS, "");
- int i;
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_module_t *wem = &ebpf_modules[i];
- if (wem->functions.fnct_routine)
- continue;
-
- write_chart_dimension((char *)wem->info.thread_name, (wem->enabled < NETDATA_THREAD_EBPF_STOPPING) ? 1 : 0);
- }
- ebpf_write_end_chart();
-
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, NETDATA_EBPF_LIFE_TIME, "");
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX ; i++) {
- ebpf_module_t *wem = &ebpf_modules[i];
- // Threads like VFS is slow to load and this can create an invalid number, this is the motive
- // we are also testing wem->lifetime value.
- if (wem->functions.fnct_routine)
- continue;
-
- write_chart_dimension((char *)wem->info.thread_name,
- (wem->lifetime && wem->enabled < NETDATA_THREAD_EBPF_STOPPING) ?
- (long long) (wem->lifetime - wem->running_time):
- 0) ;
- }
- ebpf_write_end_chart();
-
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, NETDATA_EBPF_LOAD_METHOD, "");
- write_chart_dimension(load_event_stat[NETDATA_EBPF_LOAD_STAT_LEGACY], (long long)plugin_statistics.legacy);
- write_chart_dimension(load_event_stat[NETDATA_EBPF_LOAD_STAT_CORE], (long long)plugin_statistics.core);
- ebpf_write_end_chart();
-
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, NETDATA_EBPF_KERNEL_MEMORY, "");
- write_chart_dimension(memlock_stat, (long long)plugin_statistics.memlock_kern);
- ebpf_write_end_chart();
-
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, NETDATA_EBPF_HASH_TABLES_LOADED, "");
- write_chart_dimension(hash_table_stat, (long long)plugin_statistics.hash_tables);
- ebpf_write_end_chart();
-
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, NETDATA_EBPF_HASH_TABLES_PER_CORE, "");
- write_chart_dimension(hash_table_core[NETDATA_EBPF_THREAD_PER_CORE], (long long)plugin_statistics.hash_percpu);
- write_chart_dimension(hash_table_core[NETDATA_EBPF_THREAD_UNIQUE], (long long)plugin_statistics.hash_unique);
- ebpf_write_end_chart();
-
- ebpf_send_global_hash_table_data();
-
- ebpf_send_hash_table_pid_data(NETDATA_EBPF_HASH_TABLES_INSERT_PID_ELEMENTS, NETDATA_EBPF_GLOBAL_TABLE_PID_TABLE_ADD);
- ebpf_send_hash_table_pid_data(NETDATA_EBPF_HASH_TABLES_REMOVE_PID_ELEMENTS, NETDATA_EBPF_GLOBAL_TABLE_PID_TABLE_DEL);
-
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_module_t *wem = &ebpf_modules[i];
- if (!wem->functions.fnct_routine)
- continue;
-
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, (char *)wem->functions.fcnt_thread_chart_name, "");
- write_chart_dimension((char *)wem->info.thread_name, (wem->enabled < NETDATA_THREAD_EBPF_STOPPING) ? 1 : 0);
- ebpf_write_end_chart();
-
- ebpf_write_begin_chart(NETDATA_MONITORING_FAMILY, (char *)wem->functions.fcnt_thread_lifetime_name, "");
- write_chart_dimension((char *)wem->info.thread_name,
- (wem->lifetime && wem->enabled < NETDATA_THREAD_EBPF_STOPPING) ?
- (long long) (wem->lifetime - wem->running_time):
- 0) ;
- ebpf_write_end_chart();
- }
-}
-
-/**
- * Update Internal Metric variable
- *
- * By default eBPF.plugin sends internal metrics for netdata, but user can
- * disable this.
- *
- * The function updates the variable used to send charts.
- */
-static void update_internal_metric_variable()
-{
- const char *s = getenv("NETDATA_INTERNALS_MONITORING");
- if (s && *s && strcmp(s, "NO") == 0)
- publish_internal_metrics = false;
-}
-
-/**
- * Create Thread Chart
- *
- * Write to standard output current values for threads charts.
- *
- * @param name is the chart name
- * @param title chart title.
- * @param units chart units
- * @param order is the chart order
- * @param update_every time used to update charts
- * @param module a module to create a specific chart.
- */
-static void ebpf_create_thread_chart(char *name,
- char *title,
- char *units,
- int order,
- int update_every,
- ebpf_module_t *module)
-{
- // common call for specific and all charts.
- ebpf_write_chart_cmd(NETDATA_MONITORING_FAMILY,
- name,
- "",
- title,
- units,
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_LINE,
- NULL,
- order,
- update_every,
- "main");
-
- if (module) {
- ebpf_write_global_dimension((char *)module->info.thread_name,
- (char *)module->info.thread_name,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
- return;
- }
-
- int i;
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_module_t *em = &ebpf_modules[i];
- if (em->functions.fnct_routine)
- continue;
-
- ebpf_write_global_dimension((char *)em->info.thread_name,
- (char *)em->info.thread_name,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
- }
-}
-
-/**
- * Create chart for Load Thread
- *
- * Write to standard output current values for load mode.
- *
- * @param update_every time used to update charts
- */
-static inline void ebpf_create_statistic_load_chart(int update_every)
-{
- ebpf_write_chart_cmd(NETDATA_MONITORING_FAMILY,
- NETDATA_EBPF_LOAD_METHOD,
- "",
- "Load info.",
- "methods",
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_LINE,
- NULL,
- NETDATA_EBPF_ORDER_STAT_LOAD_METHOD,
- update_every,
- NETDATA_EBPF_MODULE_NAME_PROCESS);
-
- ebpf_write_global_dimension(load_event_stat[NETDATA_EBPF_LOAD_STAT_LEGACY],
- load_event_stat[NETDATA_EBPF_LOAD_STAT_LEGACY],
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
-
- ebpf_write_global_dimension(load_event_stat[NETDATA_EBPF_LOAD_STAT_CORE],
- load_event_stat[NETDATA_EBPF_LOAD_STAT_CORE],
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
-}
-
-/**
- * Create chart for Kernel Memory
- *
- * Write to standard output current values for allocated memory.
- *
- * @param update_every time used to update charts
- */
-static inline void ebpf_create_statistic_kernel_memory(int update_every)
-{
- ebpf_write_chart_cmd(NETDATA_MONITORING_FAMILY,
- NETDATA_EBPF_KERNEL_MEMORY,
- "",
- "Memory allocated for hash tables.",
- "bytes",
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_LINE,
- NULL,
- NETDATA_EBPF_ORDER_STAT_KERNEL_MEMORY,
- update_every,
- NETDATA_EBPF_MODULE_NAME_PROCESS);
-
- ebpf_write_global_dimension(memlock_stat,
- memlock_stat,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
-}
-
-/**
- * Create chart Hash Table
- *
- * Write to standard output number of hash tables used with this software.
- *
- * @param update_every time used to update charts
- */
-static inline void ebpf_create_statistic_hash_tables(int update_every)
-{
- ebpf_write_chart_cmd(NETDATA_MONITORING_FAMILY,
- NETDATA_EBPF_HASH_TABLES_LOADED,
- "",
- "Number of hash tables loaded.",
- "hash tables",
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_LINE,
- NULL,
- NETDATA_EBPF_ORDER_STAT_HASH_TABLES,
- update_every,
- NETDATA_EBPF_MODULE_NAME_PROCESS);
-
- ebpf_write_global_dimension(hash_table_stat,
- hash_table_stat,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
-}
-
-/**
- * Create chart for percpu stats
- *
- * Write to standard output current values for threads.
- *
- * @param update_every time used to update charts
- */
-static inline void ebpf_create_statistic_hash_per_core(int update_every)
-{
- ebpf_write_chart_cmd(NETDATA_MONITORING_FAMILY,
- NETDATA_EBPF_HASH_TABLES_PER_CORE,
- "",
- "How threads are loading hash/array tables.",
- "threads",
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_LINE,
- NULL,
- NETDATA_EBPF_ORDER_STAT_HASH_CORE,
- update_every,
- NETDATA_EBPF_MODULE_NAME_PROCESS);
-
- ebpf_write_global_dimension(hash_table_core[NETDATA_EBPF_THREAD_PER_CORE],
- hash_table_core[NETDATA_EBPF_THREAD_PER_CORE],
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
-
- ebpf_write_global_dimension(hash_table_core[NETDATA_EBPF_THREAD_UNIQUE],
- hash_table_core[NETDATA_EBPF_THREAD_UNIQUE],
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
-}
-
-/**
- * Hash table global elements
- *
- * Write to standard output current values inside global tables.
- *
- * @param update_every time used to update charts
- */
-static void ebpf_create_statistic_hash_global_elements(int update_every)
-{
- ebpf_write_chart_cmd(NETDATA_MONITORING_FAMILY,
- NETDATA_EBPF_HASH_TABLES_GLOBAL_ELEMENTS,
- "",
- "Controllers inside global table",
- "rows",
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_LINE,
- NULL,
- NETDATA_EBPF_ORDER_STAT_HASH_GLOBAL_TABLE_TOTAL,
- update_every,
- NETDATA_EBPF_MODULE_NAME_PROCESS);
-
- int i;
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_write_global_dimension((char *)ebpf_modules[i].info.thread_name,
- (char *)ebpf_modules[i].info.thread_name,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
- }
-}
-
-/**
- * Hash table global elements
- *
- * Write to standard output current values inside global tables.
- *
- * @param update_every time used to update charts
- * @param id chart id
- * @param title chart title
- * @param order ordder chart will be shown on dashboard.
- */
-static void ebpf_create_statistic_hash_pid_table(int update_every, char *id, char *title, int order)
-{
- ebpf_write_chart_cmd(NETDATA_MONITORING_FAMILY,
- id,
- "",
- title,
- "rows",
- NETDATA_EBPF_FAMILY,
- NETDATA_EBPF_CHART_TYPE_LINE,
- NULL,
- order,
- update_every,
- NETDATA_EBPF_MODULE_NAME_PROCESS);
-
- int i;
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_module_t *wem = &ebpf_modules[i];
- if (wem->functions.apps_routine)
- ebpf_write_global_dimension((char *)wem->info.thread_name,
- (char *)wem->info.thread_name,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX]);
- }
-}
-
-/**
- * Create Statistics Charts
- *
- * Create charts that will show statistics related to eBPF plugin.
- *
- * @param update_every time used to update charts
- */
-static void ebpf_create_statistic_charts(int update_every)
-{
- static char create_charts = 1;
- update_internal_metric_variable();
- if (!publish_internal_metrics)
- return;
-
- if (!create_charts)
- return;
-
- create_charts = 0;
-
- ebpf_create_thread_chart(NETDATA_EBPF_THREADS,
- "Threads running.",
- "boolean",
- NETDATA_EBPF_ORDER_STAT_THREADS,
- update_every,
- NULL);
- /*
-#ifdef NETDATA_DEV_MODE
- EBPF_PLUGIN_FUNCTIONS(EBPF_FUNCTION_THREAD, EBPF_PLUGIN_THREAD_FUNCTION_DESCRIPTION);
-#endif
- */
-
- ebpf_create_thread_chart(NETDATA_EBPF_LIFE_TIME,
- "Time remaining for thread.",
- "seconds",
- NETDATA_EBPF_ORDER_STAT_LIFE_TIME,
- update_every,
- NULL);
- /*
-#ifdef NETDATA_DEV_MODE
- EBPF_PLUGIN_FUNCTIONS(EBPF_FUNCTION_THREAD, EBPF_PLUGIN_THREAD_FUNCTION_DESCRIPTION);
-#endif
- */
-
- int i,j;
- char name[256];
- for (i = 0, j = NETDATA_EBPF_ORDER_FUNCTION_PER_THREAD; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- ebpf_module_t *em = &ebpf_modules[i];
- if (!em->functions.fnct_routine)
- continue;
-
- em->functions.order_thread_chart = j;
- snprintfz(name, sizeof(name) - 1, "%s_%s", NETDATA_EBPF_THREADS, em->info.thread_name);
- em->functions.fcnt_thread_chart_name = strdupz(name);
- ebpf_create_thread_chart(name,
- "Threads running.",
- "boolean",
- j++,
- update_every,
- em);
-#ifdef NETDATA_DEV_MODE
- EBPF_PLUGIN_FUNCTIONS(em->functions.fcnt_name, em->functions.fcnt_desc);
-#endif
-
- em->functions.order_thread_lifetime = j;
- snprintfz(name, sizeof(name) - 1, "%s_%s", NETDATA_EBPF_LIFE_TIME, em->info.thread_name);
- em->functions.fcnt_thread_lifetime_name = strdupz(name);
- ebpf_create_thread_chart(name,
- "Time remaining for thread.",
- "seconds",
- j++,
- update_every,
- em);
-#ifdef NETDATA_DEV_MODE
- EBPF_PLUGIN_FUNCTIONS(em->functions.fcnt_name, em->functions.fcnt_desc);
-#endif
- }
-
- ebpf_create_statistic_load_chart(update_every);
-
- ebpf_create_statistic_kernel_memory(update_every);
-
- ebpf_create_statistic_hash_tables(update_every);
-
- ebpf_create_statistic_hash_per_core(update_every);
-
- ebpf_create_statistic_hash_global_elements(update_every);
-
- ebpf_create_statistic_hash_pid_table(update_every,
- NETDATA_EBPF_HASH_TABLES_INSERT_PID_ELEMENTS,
- "Elements inserted into PID table",
- NETDATA_EBPF_ORDER_STAT_HASH_PID_TABLE_ADDED);
-
- ebpf_create_statistic_hash_pid_table(update_every,
- NETDATA_EBPF_HASH_TABLES_REMOVE_PID_ELEMENTS,
- "Elements removed from PID table",
- NETDATA_EBPF_ORDER_STAT_HASH_PID_TABLE_REMOVED);
-
- fflush(stdout);
-}
-
-/*****************************************************************
- *
- * COLLECTOR ENTRY POINT
- *
- *****************************************************************/
-
-/**
- * Update PID file
- *
- * Update the content of PID file
- *
- * @param filename is the full name of the file.
- * @param pid that identifies the process
- */
-static void ebpf_update_pid_file(char *filename, pid_t pid)
-{
- FILE *fp = fopen(filename, "w");
- if (!fp)
- return;
-
- fprintf(fp, "%d", pid);
- fclose(fp);
-}
-
-/**
- * Get Process Name
- *
- * Get process name from /proc/PID/status
- *
- * @param pid that identifies the process
- */
-static char *ebpf_get_process_name(pid_t pid)
-{
- char *name = NULL;
- char filename[FILENAME_MAX + 1];
- snprintfz(filename, FILENAME_MAX, "/proc/%d/status", pid);
-
- procfile *ff = procfile_open(filename, " \t", PROCFILE_FLAG_DEFAULT);
- if(unlikely(!ff)) {
- netdata_log_error("Cannot open %s", filename);
- return name;
- }
-
- ff = procfile_readall(ff);
- if(unlikely(!ff))
- return name;
-
- unsigned long i, lines = procfile_lines(ff);
- for(i = 0; i < lines ; i++) {
- char *cmp = procfile_lineword(ff, i, 0);
- if (!strcmp(cmp, "Name:")) {
- name = strdupz(procfile_lineword(ff, i, 1));
- break;
- }
- }
-
- procfile_close(ff);
-
- return name;
-}
-
-/**
- * Read Previous PID
- *
- * @param filename is the full name of the file.
- *
- * @return It returns the PID used during previous execution on success or 0 otherwise
- */
-static pid_t ebpf_read_previous_pid(char *filename)
-{
- FILE *fp = fopen(filename, "r");
- if (!fp)
- return 0;
-
- char buffer[64];
- size_t length = fread(buffer, sizeof(*buffer), 63, fp);
- pid_t old_pid = 0;
- if (length) {
- if (length > 63)
- length = 63;
-
- buffer[length] = '\0';
- old_pid = (pid_t) str2uint32_t(buffer, NULL);
- }
- fclose(fp);
-
- return old_pid;
-}
-
-/**
- * Kill previous process
- *
- * Kill previous process whether it was not closed.
- *
- * @param filename is the full name of the file.
- * @param pid that identifies the process
- */
-static void ebpf_kill_previous_process(char *filename, pid_t pid)
-{
- pid_t old_pid = ebpf_read_previous_pid(filename);
- if (!old_pid)
- return;
-
- // Process is not running
- char *prev_name = ebpf_get_process_name(old_pid);
- if (!prev_name)
- return;
-
- char *current_name = ebpf_get_process_name(pid);
-
- if (!strcmp(prev_name, current_name))
- kill(old_pid, SIGKILL);
-
- freez(prev_name);
- freez(current_name);
-
- // wait few microseconds before start new plugin
- sleep_usec(USEC_PER_MS * 300);
-}
-
-/**
- * PID file
- *
- * Write the filename for PID inside the given vector.
- *
- * @param filename vector where we will store the name.
- * @param length number of bytes available in filename vector
- */
-void ebpf_pid_file(char *filename, size_t length)
-{
- snprintfz(filename, length, "%s/var/run/ebpf.pid", netdata_configured_host_prefix);
-}
-
-/**
- * Manage PID
- *
- * This function kills another instance of eBPF whether it is necessary and update the file content.
- *
- * @param pid that identifies the process
- */
-static void ebpf_manage_pid(pid_t pid)
-{
- char filename[FILENAME_MAX + 1];
- ebpf_pid_file(filename, FILENAME_MAX);
-
- ebpf_kill_previous_process(filename, pid);
- ebpf_update_pid_file(filename, pid);
-}
-
-/**
- * Set start routine
- *
- * Set static routine before threads to be created.
- */
- static void ebpf_set_static_routine()
- {
- int i;
- for (i = 0; ebpf_modules[i].info.thread_name; i++) {
- ebpf_threads[i].start_routine = ebpf_modules[i].functions.start_routine;
- }
- }
-
-/**
- * Entry point
- *
- * @param argc the number of arguments
- * @param argv the pointer to the arguments
- *
- * @return it returns 0 on success and another integer otherwise
- */
-int main(int argc, char **argv)
-{
- clocks_init();
- nd_log_initialize_for_external_plugins("ebpf.plugin");
-
- main_thread_id = gettid();
-
- set_global_variables();
- ebpf_parse_args(argc, argv);
- ebpf_manage_pid(getpid());
-
- if (ebpf_check_conditions())
- return 2;
-
- if (ebpf_adjust_memory_limit())
- return 3;
-
- signal(SIGINT, ebpf_stop_threads);
- signal(SIGQUIT, ebpf_stop_threads);
- signal(SIGTERM, ebpf_stop_threads);
- signal(SIGPIPE, ebpf_stop_threads);
-
- ebpf_start_pthread_variables();
-
- netdata_configured_host_prefix = getenv("NETDATA_HOST_PREFIX");
- if(verify_netdata_host_prefix(true) == -1) ebpf_exit(6);
-
- ebpf_allocate_common_vectors();
-
-#ifdef LIBBPF_MAJOR_VERSION
- libbpf_set_strict_mode(LIBBPF_STRICT_ALL);
-#endif
-
- ebpf_read_local_addresses_unsafe();
- read_local_ports("/proc/net/tcp", IPPROTO_TCP);
- read_local_ports("/proc/net/tcp6", IPPROTO_TCP);
- read_local_ports("/proc/net/udp", IPPROTO_UDP);
- read_local_ports("/proc/net/udp6", IPPROTO_UDP);
-
- ebpf_set_static_routine();
-
- cgroup_integration_thread.thread = mallocz(sizeof(netdata_thread_t));
- cgroup_integration_thread.start_routine = ebpf_cgroup_integration;
-
- netdata_thread_create(cgroup_integration_thread.thread, cgroup_integration_thread.name,
- NETDATA_THREAD_OPTION_DEFAULT, ebpf_cgroup_integration, NULL);
-
- int i;
- for (i = 0; ebpf_threads[i].name != NULL; i++) {
- struct netdata_static_thread *st = &ebpf_threads[i];
-
- ebpf_module_t *em = &ebpf_modules[i];
- em->thread = st;
- em->thread_id = i;
- if (em->enabled != NETDATA_THREAD_EBPF_NOT_RUNNING) {
- st->thread = mallocz(sizeof(netdata_thread_t));
- em->enabled = NETDATA_THREAD_EBPF_RUNNING;
- em->lifetime = EBPF_NON_FUNCTION_LIFE_TIME;
- netdata_thread_create(st->thread, st->name, NETDATA_THREAD_OPTION_DEFAULT, st->start_routine, em);
- } else {
- em->lifetime = EBPF_DEFAULT_LIFETIME;
- }
- }
-
- usec_t step = USEC_PER_SEC;
- heartbeat_t hb;
- heartbeat_init(&hb);
- int update_apps_every = (int) EBPF_CFG_UPDATE_APPS_EVERY_DEFAULT;
- int update_apps_list = update_apps_every - 1;
- int process_maps_per_core = ebpf_modules[EBPF_MODULE_PROCESS_IDX].maps_per_core;
- //Plugin will be killed when it receives a signal
- for ( ; !ebpf_plugin_exit; global_iterations_counter++) {
- (void)heartbeat_next(&hb, step);
-
- if (global_iterations_counter % EBPF_DEFAULT_UPDATE_EVERY == 0) {
- pthread_mutex_lock(&lock);
- ebpf_create_statistic_charts(EBPF_DEFAULT_UPDATE_EVERY);
-
- ebpf_send_statistic_data();
- pthread_mutex_unlock(&lock);
- fflush(stdout);
- }
-
- pthread_mutex_lock(&ebpf_exit_cleanup);
- pthread_mutex_lock(&collect_data_mutex);
- if (++update_apps_list == update_apps_every) {
- update_apps_list = 0;
- cleanup_exited_pids();
- collect_data_for_all_processes(process_pid_fd, process_maps_per_core);
-
- pthread_mutex_lock(&lock);
- ebpf_create_apps_charts(apps_groups_root_target);
- pthread_mutex_unlock(&lock);
- }
- pthread_mutex_unlock(&collect_data_mutex);
- pthread_mutex_unlock(&ebpf_exit_cleanup);
- }
-
- ebpf_stop_threads(0);
-
- return 0;
-}
diff --git a/collectors/ebpf.plugin/ebpf.d/swap.conf b/collectors/ebpf.plugin/ebpf.d/swap.conf
deleted file mode 100644
index 29d9b4204..000000000
--- a/collectors/ebpf.plugin/ebpf.d/swap.conf
+++ /dev/null
@@ -1,34 +0,0 @@
-# The `ebpf load mode` option accepts the following values :
-# `entry` : The eBPF collector only monitors calls for the functions, and does not show charts related to errors.
-# `return : In the `return` mode, the eBPF collector monitors the same kernel functions as `entry`, but also creates
-# new charts for the return of these functions, such as errors.
-#
-# The eBPF collector also creates charts for each running application through an integration with the `apps.plugin`
-# or `cgroups.plugin`.
-# If you want to disable the integration with `apps.plugin` or `cgroups.plugin` along with the above charts, change
-# the setting `apps` and `cgroups` to 'no'.
-#
-# The `ebpf type format` option accepts the following values :
-# `auto` : The eBPF collector will investigate hardware and select between the two next options.
-# `legacy`: The eBPF collector will load the legacy code. Note: This has a bigger overload.
-# `co-re` : The eBPF collector will use latest tracing method. Note: This is not available on all platforms.
-#
-# The `ebpf co-re tracing` option accepts the following values:
-# `trampoline`: This is the default mode used by the eBPF collector, due the small overhead added to host.
-# `probe` : This is the same as legacy code.
-#
-# The `maps per core` defines if hash tables will be per core or not. This option is ignored on kernels older than 4.6.
-#
-# The `lifetime` defines the time length a thread will run when it is enabled by a function.
-#
-# Uncomment lines to define specific options for thread.
-[global]
-# ebpf load mode = entry
-# apps = yes
-# cgroups = no
-# update every = 10
-# pid table size = 32768
- ebpf type format = auto
- ebpf co-re tracing = trampoline
-# maps per core = yes
- lifetime = 300
diff --git a/collectors/ebpf.plugin/ebpf.h b/collectors/ebpf.plugin/ebpf.h
deleted file mode 100644
index ad7c5a94c..000000000
--- a/collectors/ebpf.plugin/ebpf.h
+++ /dev/null
@@ -1,393 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-#ifndef NETDATA_COLLECTOR_EBPF_H
-#define NETDATA_COLLECTOR_EBPF_H 1
-
-#ifndef __FreeBSD__
-#include <linux/perf_event.h>
-#endif
-#include <stdint.h>
-#include <errno.h>
-#include <signal.h>
-#include <stdio.h>
-#include <stdlib.h>
-#include <string.h>
-#include <unistd.h>
-#include <dlfcn.h>
-
-#include <fcntl.h>
-#include <ctype.h>
-#include <dirent.h>
-
-// From libnetdata.h
-#include "libnetdata/threads/threads.h"
-#include "libnetdata/locks/locks.h"
-#include "libnetdata/avl/avl.h"
-#include "libnetdata/clocks/clocks.h"
-#include "libnetdata/config/appconfig.h"
-#include "libnetdata/ebpf/ebpf.h"
-#include "libnetdata/procfile/procfile.h"
-#include "collectors/cgroups.plugin/sys_fs_cgroup.h"
-#include "daemon/main.h"
-
-#include "ebpf_apps.h"
-#include "ebpf_functions.h"
-#include "ebpf_cgroup.h"
-
-#define NETDATA_EBPF_OLD_CONFIG_FILE "ebpf.conf"
-#define NETDATA_EBPF_CONFIG_FILE "ebpf.d.conf"
-
-#ifdef LIBBPF_MAJOR_VERSION // BTF code
-#include "includes/cachestat.skel.h"
-#include "includes/dc.skel.h"
-#include "includes/disk.skel.h"
-#include "includes/fd.skel.h"
-#include "includes/hardirq.skel.h"
-#include "includes/mdflush.skel.h"
-#include "includes/mount.skel.h"
-#include "includes/shm.skel.h"
-#include "includes/socket.skel.h"
-#include "includes/swap.skel.h"
-#include "includes/vfs.skel.h"
-
-extern struct cachestat_bpf *cachestat_bpf_obj;
-extern struct dc_bpf *dc_bpf_obj;
-extern struct disk_bpf *disk_bpf_obj;
-extern struct fd_bpf *fd_bpf_obj;
-extern struct hardirq_bpf *hardirq_bpf_obj;
-extern struct mount_bpf *mount_bpf_obj;
-extern struct mdflush_bpf *mdflush_bpf_obj;
-extern struct shm_bpf *shm_bpf_obj;
-extern struct socket_bpf *socket_bpf_obj;
-extern struct swap_bpf *bpf_obj;
-extern struct vfs_bpf *vfs_bpf_obj;
-#endif
-
-typedef struct netdata_syscall_stat {
- unsigned long bytes; // total number of bytes
- uint64_t call; // total number of calls
- uint64_t ecall; // number of calls that returned error
- struct netdata_syscall_stat *next; // Link list
-} netdata_syscall_stat_t;
-
-typedef struct netdata_publish_syscall {
- char *dimension;
- char *name;
- char *algorithm;
- unsigned long nbyte;
- unsigned long pbyte;
- uint64_t ncall;
- uint64_t pcall;
- uint64_t nerr;
- uint64_t perr;
- struct netdata_publish_syscall *next;
-} netdata_publish_syscall_t;
-
-typedef struct netdata_publish_vfs_common {
- long write;
- long read;
-
- long running;
- long zombie;
-} netdata_publish_vfs_common_t;
-
-typedef struct netdata_error_report {
- char comm[16];
- __u32 pid;
-
- int type;
- int err;
-} netdata_error_report_t;
-
-typedef struct netdata_ebpf_judy_pid {
- ARAL *pid_table;
-
- // Index for PIDs
- struct { // support for multiple indexing engines
- Pvoid_t JudyLArray; // the hash table
- RW_SPINLOCK rw_spinlock; // protect the index
- } index;
-} netdata_ebpf_judy_pid_t;
-
-typedef struct netdata_ebpf_judy_pid_stats {
- char *cmdline;
-
- // Index for Socket timestamp
- struct { // support for multiple indexing engines
- Pvoid_t JudyLArray; // the hash table
- RW_SPINLOCK rw_spinlock; // protect the index
- } socket_stats;
-} netdata_ebpf_judy_pid_stats_t;
-
-extern ebpf_module_t ebpf_modules[];
-enum ebpf_main_index {
- EBPF_MODULE_PROCESS_IDX,
- EBPF_MODULE_SOCKET_IDX,
- EBPF_MODULE_CACHESTAT_IDX,
- EBPF_MODULE_SYNC_IDX,
- EBPF_MODULE_DCSTAT_IDX,
- EBPF_MODULE_SWAP_IDX,
- EBPF_MODULE_VFS_IDX,
- EBPF_MODULE_FILESYSTEM_IDX,
- EBPF_MODULE_DISK_IDX,
- EBPF_MODULE_MOUNT_IDX,
- EBPF_MODULE_FD_IDX,
- EBPF_MODULE_HARDIRQ_IDX,
- EBPF_MODULE_SOFTIRQ_IDX,
- EBPF_MODULE_OOMKILL_IDX,
- EBPF_MODULE_SHM_IDX,
- EBPF_MODULE_MDFLUSH_IDX,
- EBPF_MODULE_FUNCTION_IDX,
- /* THREADS MUST BE INCLUDED BEFORE THIS COMMENT */
- EBPF_OPTION_ALL_CHARTS,
- EBPF_OPTION_VERSION,
- EBPF_OPTION_HELP,
- EBPF_OPTION_GLOBAL_CHART,
- EBPF_OPTION_RETURN_MODE,
- EBPF_OPTION_LEGACY,
- EBPF_OPTION_CORE,
- EBPF_OPTION_UNITTEST
-};
-
-typedef struct ebpf_tracepoint {
- bool enabled;
- char *class;
- char *event;
-} ebpf_tracepoint_t;
-
-// Copied from musl header
-#ifndef offsetof
-#if __GNUC__ > 3
-#define offsetof(type, member) __builtin_offsetof(type, member)
-#else
-#define offsetof(type, member) ((size_t)((char *)&(((type *)0)->member) - (char *)0))
-#endif
-#endif
-
-// Messages
-#define NETDATA_EBPF_DEFAULT_FNT_NOT_FOUND "Cannot find the necessary functions to monitor"
-
-// Chart definitions
-#define NETDATA_EBPF_FAMILY "ebpf"
-#define NETDATA_EBPF_IP_FAMILY "ip"
-#define NETDATA_FILESYSTEM_FAMILY "filesystem"
-#define NETDATA_EBPF_MOUNT_GLOBAL_FAMILY "mount_points"
-#define NETDATA_EBPF_CHART_TYPE_LINE "line"
-#define NETDATA_EBPF_CHART_TYPE_STACKED "stacked"
-#define NETDATA_EBPF_MEMORY_GROUP "mem"
-#define NETDATA_EBPF_SYSTEM_GROUP "system"
-#define NETDATA_SYSTEM_SWAP_SUBMENU "swap"
-#define NETDATA_SYSTEM_CGROUP_SWAP_SUBMENU "swap (eBPF)"
-#define NETDATA_SYSTEM_IPC_SHM_SUBMENU "ipc shared memory"
-#define NETDATA_MONITORING_FAMILY "netdata"
-
-// Statistics charts
-#define NETDATA_EBPF_THREADS "ebpf_threads"
-#define NETDATA_EBPF_LIFE_TIME "ebpf_life_time"
-#define NETDATA_EBPF_LOAD_METHOD "ebpf_load_methods"
-#define NETDATA_EBPF_KERNEL_MEMORY "ebpf_kernel_memory"
-#define NETDATA_EBPF_HASH_TABLES_LOADED "ebpf_hash_tables_count"
-#define NETDATA_EBPF_HASH_TABLES_PER_CORE "ebpf_hash_tables_per_core"
-#define NETDATA_EBPF_HASH_TABLES_GLOBAL_ELEMENTS "ebpf_hash_tables_global_elements"
-#define NETDATA_EBPF_HASH_TABLES_INSERT_PID_ELEMENTS "ebpf_hash_tables_insert_pid_elements"
-#define NETDATA_EBPF_HASH_TABLES_REMOVE_PID_ELEMENTS "ebpf_hash_tables_remove_pid_elements"
-
-// Log file
-#define NETDATA_DEVELOPER_LOG_FILE "developer.log"
-
-// Maximum number of processors monitored on perf events
-#define NETDATA_MAX_PROCESSOR 512
-
-// Kernel versions calculated with the formula:
-// R = MAJOR*65536 + MINOR*256 + PATCH
-#define NETDATA_KERNEL_V5_3 328448
-#define NETDATA_KERNEL_V4_15 265984
-
-#define EBPF_SYS_CLONE_IDX 11
-#define EBPF_MAX_MAPS 32
-
-#define EBPF_DEFAULT_UPDATE_EVERY 10
-
-enum ebpf_algorithms_list {
- NETDATA_EBPF_ABSOLUTE_IDX,
- NETDATA_EBPF_INCREMENTAL_IDX
-};
-
-// Threads
-void *ebpf_process_thread(void *ptr);
-void *ebpf_socket_thread(void *ptr);
-
-// Common variables
-extern pthread_mutex_t lock;
-extern pthread_mutex_t ebpf_exit_cleanup;
-extern int ebpf_nprocs;
-extern int running_on_kernel;
-extern int isrh;
-extern char *ebpf_plugin_dir;
-extern int process_pid_fd;
-
-extern pthread_mutex_t collect_data_mutex;
-
-// Common functions
-void ebpf_global_labels(netdata_syscall_stat_t *is,
- netdata_publish_syscall_t *pio,
- char **dim,
- char **name,
- int *algorithm,
- int end);
-
-void ebpf_write_chart_cmd(char *type,
- char *id,
- char *suffix,
- char *title,
- char *units,
- char *family,
- char *charttype,
- char *context,
- int order,
- int update_every,
- char *module);
-
-void ebpf_write_global_dimension(char *name, char *id, char *algorithm);
-
-void ebpf_create_global_dimension(void *ptr, int end);
-
-void ebpf_create_chart(char *type,
- char *id,
- char *title,
- char *units,
- char *family,
- char *context,
- char *charttype,
- int order,
- void (*ncd)(void *, int),
- void *move,
- int end,
- int update_every,
- char *module);
-
-void write_chart_dimension(char *dim, long long value);
-
-void write_count_chart(char *name, char *family, netdata_publish_syscall_t *move, uint32_t end);
-
-void write_err_chart(char *name, char *family, netdata_publish_syscall_t *move, int end);
-
-void write_io_chart(char *chart, char *family, char *dwrite, long long vwrite,
- char *dread, long long vread);
-
-/**
- * Create Chart labels
- *
- * @param name the label name.
- * @param value the label value.
- * @param origin the labeel source.
- */
-static inline void ebpf_create_chart_labels(char *name, char *value, int source)
-{
- fprintf(stdout, "CLABEL '%s' '%s' %d\n", name, value, source);
-}
-
-/**
- * Commit label
- *
- * Write commit label to stdout
- */
-static inline void ebpf_commit_label()
-{
- fprintf(stdout, "CLABEL_COMMIT\n");
-}
-
-/**
- * Write begin command on standard output
- *
- * @param family the chart family name
- * @param name the chart name
- * @param metric the chart suffix (used with apps and cgroups)
- */
-static inline void ebpf_write_begin_chart(char *family, char *name, char *metric)
-{
- printf("BEGIN %s.%s%s\n", family, name, metric);
-}
-
-/**
- * Write END command on stdout.
- */
-static inline void ebpf_write_end_chart()
-{
- printf("END\n");
-}
-
-int ebpf_enable_tracepoint(ebpf_tracepoint_t *tp);
-int ebpf_disable_tracepoint(ebpf_tracepoint_t *tp);
-uint32_t ebpf_enable_tracepoints(ebpf_tracepoint_t *tps);
-
-void ebpf_pid_file(char *filename, size_t length);
-
-#define EBPF_PROGRAMS_SECTION "ebpf programs"
-
-#define EBPF_COMMON_DIMENSION_PERCENTAGE "%"
-#define EBPF_PROGRAMS_SECTION "ebpf programs"
-
-#define EBPF_COMMON_DIMENSION_PERCENTAGE "%"
-#define EBPF_COMMON_DIMENSION_CALL "calls/s"
-#define EBPF_COMMON_DIMENSION_CONNECTIONS "connections/s"
-#define EBPF_COMMON_DIMENSION_BITS "kilobits/s"
-#define EBPF_COMMON_DIMENSION_BYTES "bytes/s"
-#define EBPF_COMMON_DIMENSION_DIFFERENCE "difference"
-#define EBPF_COMMON_DIMENSION_PACKETS "packets"
-#define EBPF_COMMON_DIMENSION_FILES "files"
-#define EBPF_COMMON_DIMENSION_MILLISECONDS "milliseconds"
-#define EBPF_COMMON_DIMENSION_KILLS "kills"
-
-// Common variables
-extern int debug_enabled;
-extern struct ebpf_pid_stat *ebpf_root_of_pids;
-extern ebpf_cgroup_target_t *ebpf_cgroup_pids;
-extern char *ebpf_algorithms[];
-extern struct config collector_config;
-extern netdata_ebpf_cgroup_shm_t shm_ebpf_cgroup;
-extern int shm_fd_ebpf_cgroup;
-extern sem_t *shm_sem_ebpf_cgroup;
-extern pthread_mutex_t mutex_cgroup_shm;
-extern size_t ebpf_all_pids_count;
-extern ebpf_plugin_stats_t plugin_statistics;
-#ifdef LIBBPF_MAJOR_VERSION
-extern struct btf *default_btf;
-#else
-extern void *default_btf;
-#endif
-
-// Socket functions and variables
-// Common functions
-void ebpf_process_create_apps_charts(struct ebpf_module *em, void *ptr);
-void ebpf_socket_create_apps_charts(struct ebpf_module *em, void *ptr);
-void ebpf_cachestat_create_apps_charts(struct ebpf_module *em, void *root);
-void ebpf_one_dimension_write_charts(char *family, char *chart, char *dim, long long v1);
-collected_number get_value_from_structure(char *basis, size_t offset);
-void ebpf_update_pid_table(ebpf_local_maps_t *pid, ebpf_module_t *em);
-void ebpf_write_chart_obsolete(char *type, char *id, char *suffix, char *title, char *units, char *family,
- char *charttype, char *context, int order, int update_every);
-void write_histogram_chart(char *family, char *name, const netdata_idx_t *hist, char **dimensions, uint32_t end);
-void ebpf_update_disabled_plugin_stats(ebpf_module_t *em);
-ARAL *ebpf_allocate_pid_aral(char *name, size_t size);
-void ebpf_unload_legacy_code(struct bpf_object *objects, struct bpf_link **probe_links);
-
-void ebpf_read_global_table_stats(netdata_idx_t *stats, netdata_idx_t *values, int map_fd,
- int maps_per_core, uint32_t begin, uint32_t end);
-void **ebpf_judy_insert_unsafe(PPvoid_t arr, Word_t key);
-netdata_ebpf_judy_pid_stats_t *ebpf_get_pid_from_judy_unsafe(PPvoid_t judy_array, uint32_t pid);
-
-void parse_network_viewer_section(struct config *cfg);
-void ebpf_clean_ip_structure(ebpf_network_viewer_ip_list_t **clean);
-void ebpf_clean_port_structure(ebpf_network_viewer_port_list_t **clean);
-void ebpf_read_local_addresses_unsafe();
-
-extern ebpf_filesystem_partitions_t localfs[];
-extern ebpf_sync_syscalls_t local_syscalls[];
-extern bool ebpf_plugin_exit;
-void ebpf_stop_threads(int sig);
-extern netdata_ebpf_judy_pid_t ebpf_judy_pid;
-
-#define EBPF_MAX_SYNCHRONIZATION_TIME 300
-
-#endif /* NETDATA_COLLECTOR_EBPF_H */
diff --git a/collectors/ebpf.plugin/ebpf_cgroup.h b/collectors/ebpf.plugin/ebpf_cgroup.h
deleted file mode 100644
index ba8346934..000000000
--- a/collectors/ebpf.plugin/ebpf_cgroup.h
+++ /dev/null
@@ -1,71 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-#ifndef NETDATA_EBPF_CGROUP_H
-#define NETDATA_EBPF_CGROUP_H 1
-
-#define NETDATA_EBPF_CGROUP_MAX_TRIES 3
-#define NETDATA_EBPF_CGROUP_NEXT_TRY_SEC 30
-
-#include "ebpf.h"
-#include "ebpf_apps.h"
-
-#define NETDATA_SERVICE_FAMILY "services"
-
-struct pid_on_target2 {
- int32_t pid;
- int updated;
-
- netdata_publish_swap_t swap;
- netdata_fd_stat_t fd;
- netdata_publish_vfs_t vfs;
- ebpf_process_stat_t ps;
- netdata_dcstat_pid_t dc;
- netdata_publish_shm_t shm;
- netdata_socket_t socket;
- netdata_cachestat_pid_t cachestat;
-
- struct pid_on_target2 *next;
-};
-
-enum ebpf_cgroup_flags {
- NETDATA_EBPF_CGROUP_HAS_PROCESS_CHART = 1,
- NETDATA_EBPF_CGROUP_HAS_SWAP_CHART = 1<<2,
- NETDATA_EBPF_CGROUP_HAS_SOCKET_CHART = 1<<3,
- NETDATA_EBPF_CGROUP_HAS_FD_CHART = 1<<4,
- NETDATA_EBPF_CGROUP_HAS_VFS_CHART = 1<<5,
- NETDATA_EBPF_CGROUP_HAS_OOMKILL_CHART = 1<<6,
- NETDATA_EBPF_CGROUP_HAS_CACHESTAT_CHART = 1<<7,
- NETDATA_EBPF_CGROUP_HAS_DC_CHART = 1<<8,
- NETDATA_EBPF_CGROUP_HAS_SHM_CHART = 1<<9
-};
-
-typedef struct ebpf_cgroup_target {
- char name[256]; // title
- uint32_t hash;
- uint32_t flags;
- uint32_t systemd;
- uint32_t updated;
-
- netdata_publish_swap_t publish_systemd_swap;
- netdata_fd_stat_t publish_systemd_fd;
- netdata_publish_vfs_t publish_systemd_vfs;
- ebpf_process_stat_t publish_systemd_ps;
- netdata_publish_dcstat_t publish_dc;
- int oomkill;
- netdata_publish_shm_t publish_shm;
- ebpf_socket_publish_apps_t publish_socket;
- netdata_publish_cachestat_t publish_cachestat;
-
- struct pid_on_target2 *pids;
- struct ebpf_cgroup_target *next;
-} ebpf_cgroup_target_t;
-
-void ebpf_map_cgroup_shared_memory();
-void ebpf_parse_cgroup_shm_data();
-void ebpf_create_charts_on_systemd(char *id, char *title, char *units, char *family, char *charttype, int order,
- char *algorithm, char *context, char *module, int update_every);
-void *ebpf_cgroup_integration(void *ptr);
-void ebpf_unmap_cgroup_shared_memory();
-extern int send_cgroup_chart;
-
-#endif /* NETDATA_EBPF_CGROUP_H */
diff --git a/collectors/ebpf.plugin/ebpf_functions.c b/collectors/ebpf.plugin/ebpf_functions.c
deleted file mode 100644
index 6a481ad64..000000000
--- a/collectors/ebpf.plugin/ebpf_functions.c
+++ /dev/null
@@ -1,1093 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-#include "ebpf.h"
-#include "ebpf_functions.h"
-
-/*****************************************************************
- * EBPF FUNCTION COMMON
- *****************************************************************/
-
-/**
- * Function Start thread
- *
- * Start a specific thread after user request.
- *
- * @param em The structure with thread information
- * @param period
- * @return
- */
-static int ebpf_function_start_thread(ebpf_module_t *em, int period)
-{
- struct netdata_static_thread *st = em->thread;
- // another request for thread that already ran, cleanup and restart
- if (st->thread)
- freez(st->thread);
-
- if (period <= 0)
- period = EBPF_DEFAULT_LIFETIME;
-
- st->thread = mallocz(sizeof(netdata_thread_t));
- em->enabled = NETDATA_THREAD_EBPF_FUNCTION_RUNNING;
- em->lifetime = period;
-
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("Starting thread %s with lifetime = %d", em->info.thread_name, period);
-#endif
-
- return netdata_thread_create(st->thread, st->name, NETDATA_THREAD_OPTION_DEFAULT, st->start_routine, em);
-}
-
-/*****************************************************************
- * EBPF SELECT MODULE
- *****************************************************************/
-
-/**
- * Select Module
- *
- * @param thread_name name of the thread we are looking for.
- *
- * @return it returns a pointer for the module that has thread_name on success or NULL otherwise.
-ebpf_module_t *ebpf_functions_select_module(const char *thread_name) {
- int i;
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- if (strcmp(ebpf_modules[i].info.thread_name, thread_name) == 0) {
- return &ebpf_modules[i];
- }
- }
-
- return NULL;
-}
- */
-
-/*****************************************************************
- * EBPF HELP FUNCTIONS
- *****************************************************************/
-
-/**
- * Thread Help
- *
- * Shows help with all options accepted by thread function.
- *
- * @param transaction the transaction id that Netdata sent for this function execution
-static void ebpf_function_thread_manipulation_help(const char *transaction) {
- BUFFER *wb = buffer_create(0, NULL);
- buffer_sprintf(wb, "%s",
- "ebpf.plugin / thread\n"
- "\n"
- "Function `thread` allows user to control eBPF threads.\n"
- "\n"
- "The following filters are supported:\n"
- "\n"
- " thread:NAME\n"
- " Shows information for the thread NAME. Names are listed inside `ebpf.d.conf`.\n"
- "\n"
- " enable:NAME:PERIOD\n"
- " Enable a specific thread named `NAME` to run a specific PERIOD in seconds. When PERIOD is not\n"
- " specified plugin will use the default 300 seconds\n"
- "\n"
- " disable:NAME\n"
- " Disable a sp.\n"
- "\n"
- "Filters can be combined. Each filter can be given only one time.\n"
- );
-
- pluginsd_function_result_to_stdout(transaction, HTTP_RESP_OK, "text/plain", now_realtime_sec() + 3600, wb);
-
- buffer_free(wb);
-}
-*/
-
-/*****************************************************************
- * EBPF ERROR FUNCTIONS
- *****************************************************************/
-
-/**
- * Function error
- *
- * Show error when a wrong function is given
- *
- * @param transaction the transaction id that Netdata sent for this function execution
- * @param code the error code to show with the message.
- * @param msg the error message
- */
-static void ebpf_function_error(const char *transaction, int code, const char *msg) {
- pluginsd_function_json_error_to_stdout(transaction, code, msg);
-}
-
-/*****************************************************************
- * EBPF THREAD FUNCTION
- *****************************************************************/
-
-/**
- * Function: thread
- *
- * Enable a specific thread.
- *
- * @param transaction the transaction id that Netdata sent for this function execution
- * @param function function name and arguments given to thread.
- * @param line_buffer buffer used to parse args
- * @param line_max Number of arguments given
- * @param timeout The function timeout
- * @param em The structure with thread information
-static void ebpf_function_thread_manipulation(const char *transaction,
- char *function __maybe_unused,
- char *line_buffer __maybe_unused,
- int line_max __maybe_unused,
- int timeout __maybe_unused,
- ebpf_module_t *em)
-{
- char *words[PLUGINSD_MAX_WORDS] = { NULL };
- char message[512];
- uint32_t show_specific_thread = 0;
- size_t num_words = quoted_strings_splitter_pluginsd(function, words, PLUGINSD_MAX_WORDS);
- for(int i = 1; i < PLUGINSD_MAX_WORDS ;i++) {
- const char *keyword = get_word(words, num_words, i);
- if (!keyword)
- break;
-
- ebpf_module_t *lem;
- if(strncmp(keyword, EBPF_THREADS_ENABLE_CATEGORY, sizeof(EBPF_THREADS_ENABLE_CATEGORY) -1) == 0) {
- char thread_name[128];
- int period = -1;
- const char *name = &keyword[sizeof(EBPF_THREADS_ENABLE_CATEGORY) - 1];
- char *separator = strchr(name, ':');
- if (separator) {
- strncpyz(thread_name, name, separator - name);
- period = str2i(++separator);
- } else {
- strncpyz(thread_name, name, strlen(name));
- }
-
- lem = ebpf_functions_select_module(thread_name);
- if (!lem) {
- snprintfz(message, sizeof(message) - 1, "%s%s", EBPF_PLUGIN_THREAD_FUNCTION_ERROR_THREAD_NOT_FOUND, name);
- ebpf_function_error(transaction, HTTP_RESP_NOT_FOUND, message);
- return;
- }
-
- pthread_mutex_lock(&ebpf_exit_cleanup);
- if (lem->enabled > NETDATA_THREAD_EBPF_FUNCTION_RUNNING) {
- // Load configuration again
- ebpf_update_module(lem, default_btf, running_on_kernel, isrh);
-
- if (ebpf_function_start_thread(lem, period)) {
- ebpf_function_error(transaction,
- HTTP_RESP_INTERNAL_SERVER_ERROR,
- "Cannot start thread.");
- return;
- }
- } else {
- lem->running_time = 0;
- if (period > 0) // user is modifying period to run
- lem->lifetime = period;
-#ifdef NETDATA_INTERNAL_CHECKS
- netdata_log_info("Thread %s had lifetime updated for %d", thread_name, period);
-#endif
- }
- pthread_mutex_unlock(&ebpf_exit_cleanup);
- } else if(strncmp(keyword, EBPF_THREADS_DISABLE_CATEGORY, sizeof(EBPF_THREADS_DISABLE_CATEGORY) -1) == 0) {
- const char *name = &keyword[sizeof(EBPF_THREADS_DISABLE_CATEGORY) - 1];
- lem = ebpf_functions_select_module(name);
- if (!lem) {
- snprintfz(message, sizeof(message) - 1, "%s%s", EBPF_PLUGIN_THREAD_FUNCTION_ERROR_THREAD_NOT_FOUND, name);
- ebpf_function_error(transaction, HTTP_RESP_NOT_FOUND, message);
- return;
- }
-
- pthread_mutex_lock(&ebpf_exit_cleanup);
- if (lem->enabled < NETDATA_THREAD_EBPF_STOPPING && lem->thread->thread) {
- lem->lifetime = 0;
- lem->running_time = lem->update_every;
- netdata_thread_cancel(*lem->thread->thread);
- }
- pthread_mutex_unlock(&ebpf_exit_cleanup);
- } else if(strncmp(keyword, EBPF_THREADS_SELECT_THREAD, sizeof(EBPF_THREADS_SELECT_THREAD) -1) == 0) {
- const char *name = &keyword[sizeof(EBPF_THREADS_SELECT_THREAD) - 1];
- lem = ebpf_functions_select_module(name);
- if (!lem) {
- snprintfz(message, sizeof(message) - 1, "%s%s", EBPF_PLUGIN_THREAD_FUNCTION_ERROR_THREAD_NOT_FOUND, name);
- ebpf_function_error(transaction, HTTP_RESP_NOT_FOUND, message);
- return;
- }
-
- show_specific_thread |= 1<<lem->thread_id;
- } else if(strncmp(keyword, "help", 4) == 0) {
- ebpf_function_thread_manipulation_help(transaction);
- return;
- }
- }
-
- time_t expires = now_realtime_sec() + em->update_every;
-
- BUFFER *wb = buffer_create(PLUGINSD_LINE_MAX, NULL);
- buffer_json_initialize(wb, "\"", "\"", 0, true, BUFFER_JSON_OPTIONS_NEWLINE_ON_ARRAY_ITEMS);
- buffer_json_member_add_uint64(wb, "status", HTTP_RESP_OK);
- buffer_json_member_add_string(wb, "type", "table");
- buffer_json_member_add_time_t(wb, "update_every", em->update_every);
- buffer_json_member_add_string(wb, "help", EBPF_PLUGIN_THREAD_FUNCTION_DESCRIPTION);
-
- // Collect data
- buffer_json_member_add_array(wb, "data");
- int i;
- for (i = 0; i < EBPF_MODULE_FUNCTION_IDX; i++) {
- if (show_specific_thread && !(show_specific_thread & 1<<i))
- continue;
-
- ebpf_module_t *wem = &ebpf_modules[i];
- buffer_json_add_array_item_array(wb);
-
- // IMPORTANT!
- // THE ORDER SHOULD BE THE SAME WITH THE FIELDS!
-
- // thread name
- buffer_json_add_array_item_string(wb, wem->info.thread_name);
-
- // description
- buffer_json_add_array_item_string(wb, wem->info.thread_description);
- // Either it is not running or received a disabled signal and it is stopping.
- if (wem->enabled > NETDATA_THREAD_EBPF_FUNCTION_RUNNING ||
- (!wem->lifetime && (int)wem->running_time == wem->update_every)) {
- // status
- buffer_json_add_array_item_string(wb, EBPF_THREAD_STATUS_STOPPED);
-
- // Time remaining
- buffer_json_add_array_item_uint64(wb, 0);
-
- // action
- buffer_json_add_array_item_string(wb, "NULL");
- } else {
- // status
- buffer_json_add_array_item_string(wb, EBPF_THREAD_STATUS_RUNNING);
-
- // Time remaining
- buffer_json_add_array_item_uint64(wb, (wem->lifetime) ? (wem->lifetime - wem->running_time) : 0);
-
- // action
- buffer_json_add_array_item_string(wb, "Enabled/Disabled");
- }
-
- buffer_json_array_close(wb);
- }
-
- buffer_json_array_close(wb); // data
-
- buffer_json_member_add_object(wb, "columns");
- {
- int fields_id = 0;
-
- // IMPORTANT!
- // THE ORDER SHOULD BE THE SAME WITH THE VALUES!
- buffer_rrdf_table_add_field(wb, fields_id++, "Thread", "Thread Name", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY | RRDF_FIELD_OPTS_UNIQUE_KEY, NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Description", "Thread Desc", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Status", "Thread Status", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Time", "Time Remaining", RRDF_FIELD_TYPE_INTEGER,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NUMBER, 0, NULL,
- NAN, RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_NONE, NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Action", "Thread Action", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
- }
- buffer_json_object_close(wb); // columns
-
- buffer_json_member_add_string(wb, "default_sort_column", "Thread");
-
- buffer_json_member_add_object(wb, "charts");
- {
- // Threads
- buffer_json_member_add_object(wb, "eBPFThreads");
- {
- buffer_json_member_add_string(wb, "name", "Threads");
- buffer_json_member_add_string(wb, "type", "line");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Threads");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // Life Time
- buffer_json_member_add_object(wb, "eBPFLifeTime");
- {
- buffer_json_member_add_string(wb, "name", "LifeTime");
- buffer_json_member_add_string(wb, "type", "line");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Threads");
- buffer_json_add_array_item_string(wb, "Time");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
- }
- buffer_json_object_close(wb); // charts
-
- // Do we use only on fields that can be groupped?
- buffer_json_member_add_object(wb, "group_by");
- {
- // group by Status
- buffer_json_member_add_object(wb, "Status");
- {
- buffer_json_member_add_string(wb, "name", "Thread status");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Status");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
- }
- buffer_json_object_close(wb); // group_by
-
- buffer_json_member_add_time_t(wb, "expires", expires);
- buffer_json_finalize(wb);
-
- // Lock necessary to avoid race condition
- pluginsd_function_result_to_stdout(transaction, HTTP_RESP_OK, "application/json", expires, wb);
-
- buffer_free(wb);
-}
- */
-
-/*****************************************************************
- * EBPF SOCKET FUNCTION
- *****************************************************************/
-
-/**
- * Thread Help
- *
- * Shows help with all options accepted by thread function.
- *
- * @param transaction the transaction id that Netdata sent for this function execution
-*/
-static void ebpf_function_socket_help(const char *transaction) {
- pluginsd_function_result_begin_to_stdout(transaction, HTTP_RESP_OK, "text/plain", now_realtime_sec() + 3600);
- fprintf(stdout, "%s",
- "ebpf.plugin / socket\n"
- "\n"
- "Function `socket` display information for all open sockets during ebpf.plugin runtime.\n"
- "During thread runtime the plugin is always collecting data, but when an option is modified, the plugin\n"
- "resets completely the previous table and can show a clean data for the first request before to bring the\n"
- "modified request.\n"
- "\n"
- "The following filters are supported:\n"
- "\n"
- " family:FAMILY\n"
- " Shows information for the FAMILY specified. Option accepts IPV4, IPV6 and all, that is the default.\n"
- "\n"
- " period:PERIOD\n"
- " Enable socket to run a specific PERIOD in seconds. When PERIOD is not\n"
- " specified plugin will use the default 300 seconds\n"
- "\n"
- " resolve:BOOL\n"
- " Resolve service name, default value is YES.\n"
- "\n"
- " range:CIDR\n"
- " Show sockets that have only a specific destination. Default all addresses.\n"
- "\n"
- " port:range\n"
- " Show sockets that have only a specific destination.\n"
- "\n"
- " reset\n"
- " Send a reset to collector. When a collector receives this command, it uses everything defined in configuration file.\n"
- "\n"
- " interfaces\n"
- " When the collector receives this command, it read all available interfaces on host.\n"
- "\n"
- "Filters can be combined. Each filter can be given only one time. Default all ports\n"
- );
- pluginsd_function_result_end_to_stdout();
- fflush(stdout);
-}
-
-/**
- * Fill Fake socket
- *
- * Fill socket with an invalid request.
- *
- * @param fake_values is the structure where we are storing the value.
- */
-static inline void ebpf_socket_fill_fake_socket(netdata_socket_plus_t *fake_values)
-{
- snprintfz(fake_values->socket_string.src_ip, INET6_ADDRSTRLEN, "%s", "127.0.0.1");
- snprintfz(fake_values->socket_string.dst_ip, INET6_ADDRSTRLEN, "%s", "127.0.0.1");
- fake_values->pid = getpid();
- //fake_values->socket_string.src_port = 0;
- fake_values->socket_string.dst_port[0] = 0;
- snprintfz(fake_values->socket_string.dst_ip, NI_MAXSERV, "%s", "none");
- fake_values->data.family = AF_INET;
- fake_values->data.protocol = AF_UNSPEC;
-}
-
-/**
- * Fill function buffer
- *
- * Fill buffer with data to be shown on cloud.
- *
- * @param wb buffer where we store data.
- * @param values data read from hash table
- * @param name the process name
- */
-static void ebpf_fill_function_buffer(BUFFER *wb, netdata_socket_plus_t *values, char *name)
-{
- buffer_json_add_array_item_array(wb);
-
- // IMPORTANT!
- // THE ORDER SHOULD BE THE SAME WITH THE FIELDS!
-
- // PID
- buffer_json_add_array_item_uint64(wb, (uint64_t)values->pid);
-
- // NAME
- buffer_json_add_array_item_string(wb, (name) ? name : "not identified");
-
- // Origin
- buffer_json_add_array_item_string(wb, (values->data.external_origin) ? "incoming" : "outgoing");
-
- // Source IP
- buffer_json_add_array_item_string(wb, values->socket_string.src_ip);
-
- // SRC Port
- //buffer_json_add_array_item_uint64(wb, (uint64_t) values->socket_string.src_port);
-
- // Destination IP
- buffer_json_add_array_item_string(wb, values->socket_string.dst_ip);
-
- // DST Port
- buffer_json_add_array_item_string(wb, values->socket_string.dst_port);
-
- uint64_t connections;
- if (values->data.protocol == IPPROTO_TCP) {
- // Protocol
- buffer_json_add_array_item_string(wb, "TCP");
-
- // Bytes received
- buffer_json_add_array_item_uint64(wb, (uint64_t) values->data.tcp.tcp_bytes_received);
-
- // Bytes sent
- buffer_json_add_array_item_uint64(wb, (uint64_t) values->data.tcp.tcp_bytes_sent);
-
- // Connections
- connections = values->data.tcp.ipv4_connect + values->data.tcp.ipv6_connect;
- } else if (values->data.protocol == IPPROTO_UDP) {
- // Protocol
- buffer_json_add_array_item_string(wb, "UDP");
-
- // Bytes received
- buffer_json_add_array_item_uint64(wb, (uint64_t) values->data.udp.udp_bytes_received);
-
- // Bytes sent
- buffer_json_add_array_item_uint64(wb, (uint64_t) values->data.udp.udp_bytes_sent);
-
- // Connections
- connections = values->data.udp.call_udp_sent + values->data.udp.call_udp_received;
- } else {
- // Protocol
- buffer_json_add_array_item_string(wb, "UNSPEC");
-
- // Bytes received
- buffer_json_add_array_item_uint64(wb, 0);
-
- // Bytes sent
- buffer_json_add_array_item_uint64(wb, 0);
-
- connections = 1;
- }
-
- // Connections
- if (values->flags & NETDATA_SOCKET_FLAGS_ALREADY_OPEN) {
- connections++;
- } else if (!connections) {
- // If no connections, this means that we lost when connection was opened
- values->flags |= NETDATA_SOCKET_FLAGS_ALREADY_OPEN;
- connections++;
- }
- buffer_json_add_array_item_uint64(wb, connections);
-
- buffer_json_array_close(wb);
-}
-
-/**
- * Clean Judy array unsafe
- *
- * Clean all Judy Array allocated to show table when a function is called.
- * Before to call this function it is necessary to lock `ebpf_judy_pid.index.rw_spinlock`.
- **/
-static void ebpf_socket_clean_judy_array_unsafe()
-{
- if (!ebpf_judy_pid.index.JudyLArray)
- return;
-
- Pvoid_t *pid_value, *socket_value;
- Word_t local_pid = 0, local_socket = 0;
- bool first_pid = true, first_socket = true;
- while ((pid_value = JudyLFirstThenNext(ebpf_judy_pid.index.JudyLArray, &local_pid, &first_pid))) {
- netdata_ebpf_judy_pid_stats_t *pid_ptr = (netdata_ebpf_judy_pid_stats_t *)*pid_value;
- rw_spinlock_write_lock(&pid_ptr->socket_stats.rw_spinlock);
- if (pid_ptr->socket_stats.JudyLArray) {
- while ((socket_value = JudyLFirstThenNext(pid_ptr->socket_stats.JudyLArray, &local_socket, &first_socket))) {
- netdata_socket_plus_t *socket_clean = *socket_value;
- aral_freez(aral_socket_table, socket_clean);
- }
- JudyLFreeArray(&pid_ptr->socket_stats.JudyLArray, PJE0);
- pid_ptr->socket_stats.JudyLArray = NULL;
- }
- rw_spinlock_write_unlock(&pid_ptr->socket_stats.rw_spinlock);
- }
-}
-
-/**
- * Fill function buffer unsafe
- *
- * Fill the function buffer with socket information. Before to call this function it is necessary to lock
- * ebpf_judy_pid.index.rw_spinlock
- *
- * @param buf buffer used to store data to be shown by function.
- *
- * @return it returns 0 on success and -1 otherwise.
- */
-static void ebpf_socket_fill_function_buffer_unsafe(BUFFER *buf)
-{
- int counter = 0;
-
- Pvoid_t *pid_value, *socket_value;
- Word_t local_pid = 0;
- bool first_pid = true;
- while ((pid_value = JudyLFirstThenNext(ebpf_judy_pid.index.JudyLArray, &local_pid, &first_pid))) {
- netdata_ebpf_judy_pid_stats_t *pid_ptr = (netdata_ebpf_judy_pid_stats_t *)*pid_value;
- bool first_socket = true;
- Word_t local_timestamp = 0;
- rw_spinlock_read_lock(&pid_ptr->socket_stats.rw_spinlock);
- if (pid_ptr->socket_stats.JudyLArray) {
- while ((socket_value = JudyLFirstThenNext(pid_ptr->socket_stats.JudyLArray, &local_timestamp, &first_socket))) {
- netdata_socket_plus_t *values = (netdata_socket_plus_t *)*socket_value;
- ebpf_fill_function_buffer(buf, values, pid_ptr->cmdline);
- }
- counter++;
- }
- rw_spinlock_read_unlock(&pid_ptr->socket_stats.rw_spinlock);
- }
-
- if (!counter) {
- netdata_socket_plus_t fake_values = { };
- ebpf_socket_fill_fake_socket(&fake_values);
- ebpf_fill_function_buffer(buf, &fake_values, NULL);
- }
-}
-
-/**
- * Socket read hash
- *
- * This is the thread callback.
- * This thread is necessary, because we cannot freeze the whole plugin to read the data on very busy socket.
- *
- * @param buf the buffer to store data;
- * @param em the module main structure.
- *
- * @return It always returns NULL.
- */
-void ebpf_socket_read_open_connections(BUFFER *buf, struct ebpf_module *em)
-{
- // thread was not initialized or Array was reset
- rw_spinlock_read_lock(&ebpf_judy_pid.index.rw_spinlock);
- if (!em->maps || (em->maps[NETDATA_SOCKET_OPEN_SOCKET].map_fd == ND_EBPF_MAP_FD_NOT_INITIALIZED) ||
- !ebpf_judy_pid.index.JudyLArray){
- netdata_socket_plus_t fake_values = { };
-
- ebpf_socket_fill_fake_socket(&fake_values);
-
- ebpf_fill_function_buffer(buf, &fake_values, NULL);
- rw_spinlock_read_unlock(&ebpf_judy_pid.index.rw_spinlock);
- return;
- }
-
- rw_spinlock_read_lock(&network_viewer_opt.rw_spinlock);
- ebpf_socket_fill_function_buffer_unsafe(buf);
- rw_spinlock_read_unlock(&network_viewer_opt.rw_spinlock);
- rw_spinlock_read_unlock(&ebpf_judy_pid.index.rw_spinlock);
-}
-
-/**
- * Function: Socket
- *
- * Show information for sockets stored in hash tables.
- *
- * @param transaction the transaction id that Netdata sent for this function execution
- * @param function function name and arguments given to thread.
- * @param timeout The function timeout
- * @param cancelled Variable used to store function status.
- */
-static void ebpf_function_socket_manipulation(const char *transaction,
- char *function __maybe_unused,
- int timeout __maybe_unused,
- bool *cancelled __maybe_unused)
-{
- UNUSED(timeout);
- ebpf_module_t *em = &ebpf_modules[EBPF_MODULE_SOCKET_IDX];
-
- char *words[PLUGINSD_MAX_WORDS] = {NULL};
- size_t num_words = quoted_strings_splitter_pluginsd(function, words, PLUGINSD_MAX_WORDS);
- const char *name;
- int period = -1;
- rw_spinlock_write_lock(&ebpf_judy_pid.index.rw_spinlock);
- network_viewer_opt.enabled = CONFIG_BOOLEAN_YES;
- uint32_t previous;
-
- for (int i = 1; i < PLUGINSD_MAX_WORDS; i++) {
- const char *keyword = get_word(words, num_words, i);
- if (!keyword)
- break;
-
- if (strncmp(keyword, EBPF_FUNCTION_SOCKET_FAMILY, sizeof(EBPF_FUNCTION_SOCKET_FAMILY) - 1) == 0) {
- name = &keyword[sizeof(EBPF_FUNCTION_SOCKET_FAMILY) - 1];
- previous = network_viewer_opt.family;
- uint32_t family = AF_UNSPEC;
- if (!strcmp(name, "IPV4"))
- family = AF_INET;
- else if (!strcmp(name, "IPV6"))
- family = AF_INET6;
-
- if (family != previous) {
- rw_spinlock_write_lock(&network_viewer_opt.rw_spinlock);
- network_viewer_opt.family = family;
- rw_spinlock_write_unlock(&network_viewer_opt.rw_spinlock);
- ebpf_socket_clean_judy_array_unsafe();
- }
- } else if (strncmp(keyword, EBPF_FUNCTION_SOCKET_PERIOD, sizeof(EBPF_FUNCTION_SOCKET_PERIOD) - 1) == 0) {
- name = &keyword[sizeof(EBPF_FUNCTION_SOCKET_PERIOD) - 1];
- pthread_mutex_lock(&ebpf_exit_cleanup);
- period = str2i(name);
- if (period > 0) {
- em->lifetime = period;
- } else
- em->lifetime = EBPF_NON_FUNCTION_LIFE_TIME;
-
-#ifdef NETDATA_DEV_MODE
- collector_info("Lifetime modified for %u", em->lifetime);
-#endif
- pthread_mutex_unlock(&ebpf_exit_cleanup);
- } else if (strncmp(keyword, EBPF_FUNCTION_SOCKET_RESOLVE, sizeof(EBPF_FUNCTION_SOCKET_RESOLVE) - 1) == 0) {
- previous = network_viewer_opt.service_resolution_enabled;
- uint32_t resolution;
- name = &keyword[sizeof(EBPF_FUNCTION_SOCKET_RESOLVE) - 1];
- resolution = (!strcasecmp(name, "YES")) ? CONFIG_BOOLEAN_YES : CONFIG_BOOLEAN_NO;
-
- if (previous != resolution) {
- rw_spinlock_write_lock(&network_viewer_opt.rw_spinlock);
- network_viewer_opt.service_resolution_enabled = resolution;
- rw_spinlock_write_unlock(&network_viewer_opt.rw_spinlock);
-
- ebpf_socket_clean_judy_array_unsafe();
- }
- } else if (strncmp(keyword, EBPF_FUNCTION_SOCKET_RANGE, sizeof(EBPF_FUNCTION_SOCKET_RANGE) - 1) == 0) {
- name = &keyword[sizeof(EBPF_FUNCTION_SOCKET_RANGE) - 1];
- rw_spinlock_write_lock(&network_viewer_opt.rw_spinlock);
- ebpf_clean_ip_structure(&network_viewer_opt.included_ips);
- ebpf_clean_ip_structure(&network_viewer_opt.excluded_ips);
- ebpf_parse_ips_unsafe((char *)name);
- rw_spinlock_write_unlock(&network_viewer_opt.rw_spinlock);
-
- ebpf_socket_clean_judy_array_unsafe();
- } else if (strncmp(keyword, EBPF_FUNCTION_SOCKET_PORT, sizeof(EBPF_FUNCTION_SOCKET_PORT) - 1) == 0) {
- name = &keyword[sizeof(EBPF_FUNCTION_SOCKET_PORT) - 1];
- rw_spinlock_write_lock(&network_viewer_opt.rw_spinlock);
- ebpf_clean_port_structure(&network_viewer_opt.included_port);
- ebpf_clean_port_structure(&network_viewer_opt.excluded_port);
- ebpf_parse_ports((char *)name);
- rw_spinlock_write_unlock(&network_viewer_opt.rw_spinlock);
-
- ebpf_socket_clean_judy_array_unsafe();
- } else if (strncmp(keyword, EBPF_FUNCTION_SOCKET_RESET, sizeof(EBPF_FUNCTION_SOCKET_RESET) - 1) == 0) {
- rw_spinlock_write_lock(&network_viewer_opt.rw_spinlock);
- ebpf_clean_port_structure(&network_viewer_opt.included_port);
- ebpf_clean_port_structure(&network_viewer_opt.excluded_port);
-
- ebpf_clean_ip_structure(&network_viewer_opt.included_ips);
- ebpf_clean_ip_structure(&network_viewer_opt.excluded_ips);
- ebpf_clean_ip_structure(&network_viewer_opt.ipv4_local_ip);
- ebpf_clean_ip_structure(&network_viewer_opt.ipv6_local_ip);
-
- parse_network_viewer_section(&socket_config);
- ebpf_read_local_addresses_unsafe();
- network_viewer_opt.enabled = CONFIG_BOOLEAN_YES;
- rw_spinlock_write_unlock(&network_viewer_opt.rw_spinlock);
- } else if (strncmp(keyword, EBPF_FUNCTION_SOCKET_INTERFACES, sizeof(EBPF_FUNCTION_SOCKET_INTERFACES) - 1) == 0) {
- rw_spinlock_write_lock(&network_viewer_opt.rw_spinlock);
- ebpf_read_local_addresses_unsafe();
- rw_spinlock_write_unlock(&network_viewer_opt.rw_spinlock);
- } else if (strncmp(keyword, "help", 4) == 0) {
- ebpf_function_socket_help(transaction);
- rw_spinlock_write_unlock(&ebpf_judy_pid.index.rw_spinlock);
- return;
- }
- }
- rw_spinlock_write_unlock(&ebpf_judy_pid.index.rw_spinlock);
-
- pthread_mutex_lock(&ebpf_exit_cleanup);
- if (em->enabled > NETDATA_THREAD_EBPF_FUNCTION_RUNNING) {
- // Cleanup when we already had a thread running
- rw_spinlock_write_lock(&ebpf_judy_pid.index.rw_spinlock);
- ebpf_socket_clean_judy_array_unsafe();
- rw_spinlock_write_unlock(&ebpf_judy_pid.index.rw_spinlock);
-
- if (ebpf_function_start_thread(em, period)) {
- ebpf_function_error(transaction,
- HTTP_RESP_INTERNAL_SERVER_ERROR,
- "Cannot start thread.");
- pthread_mutex_unlock(&ebpf_exit_cleanup);
- return;
- }
- } else {
- if (period < 0 && em->lifetime < EBPF_NON_FUNCTION_LIFE_TIME) {
- em->lifetime = EBPF_NON_FUNCTION_LIFE_TIME;
- }
- }
- pthread_mutex_unlock(&ebpf_exit_cleanup);
-
- time_t expires = now_realtime_sec() + em->update_every;
-
- BUFFER *wb = buffer_create(PLUGINSD_LINE_MAX, NULL);
- buffer_json_initialize(wb, "\"", "\"", 0, true, false);
- buffer_json_member_add_uint64(wb, "status", HTTP_RESP_OK);
- buffer_json_member_add_string(wb, "type", "table");
- buffer_json_member_add_time_t(wb, "update_every", em->update_every);
- buffer_json_member_add_string(wb, "help", EBPF_PLUGIN_SOCKET_FUNCTION_DESCRIPTION);
-
- // Collect data
- buffer_json_member_add_array(wb, "data");
- ebpf_socket_read_open_connections(wb, em);
- buffer_json_array_close(wb); // data
-
- buffer_json_member_add_object(wb, "columns");
- {
- int fields_id = 0;
-
- // IMPORTANT!
- // THE ORDER SHOULD BE THE SAME WITH THE VALUES!
- buffer_rrdf_table_add_field(wb, fields_id++, "PID", "Process ID", RRDF_FIELD_TYPE_INTEGER,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NUMBER, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY,
- NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Process Name", "Process Name", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Origin", "The connection origin.", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Request from", "Request from IP", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
-
- /*
- buffer_rrdf_table_add_field(wb, fields_id++, "SRC PORT", "Source Port", RRDF_FIELD_TYPE_INTEGER,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NUMBER, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY,
- NULL);
- */
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Destination IP", "Destination IP", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Destination Port", "Destination Port", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Protocol", "Communication protocol", RRDF_FIELD_TYPE_STRING,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NONE, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY, NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Incoming Bandwidth", "Bytes received.", RRDF_FIELD_TYPE_INTEGER,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NUMBER, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY,
- NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id++, "Outgoing Bandwidth", "Bytes sent.", RRDF_FIELD_TYPE_INTEGER,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NUMBER, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY,
- NULL);
-
- buffer_rrdf_table_add_field(wb, fields_id, "Connections", "Number of calls to tcp_vX_connections and udp_sendmsg, where X is the protocol version.", RRDF_FIELD_TYPE_INTEGER,
- RRDF_FIELD_VISUAL_VALUE, RRDF_FIELD_TRANSFORM_NUMBER, 0, NULL, NAN,
- RRDF_FIELD_SORT_ASCENDING, NULL, RRDF_FIELD_SUMMARY_COUNT,
- RRDF_FIELD_FILTER_MULTISELECT,
- RRDF_FIELD_OPTS_VISIBLE | RRDF_FIELD_OPTS_STICKY,
- NULL);
- }
- buffer_json_object_close(wb); // columns
-
- buffer_json_member_add_object(wb, "charts");
- {
- // OutBound Connections
- buffer_json_member_add_object(wb, "IPInboundConn");
- {
- buffer_json_member_add_string(wb, "name", "TCP Inbound Connection");
- buffer_json_member_add_string(wb, "type", "line");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "connected_tcp");
- buffer_json_add_array_item_string(wb, "connected_udp");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // OutBound Connections
- buffer_json_member_add_object(wb, "IPTCPOutboundConn");
- {
- buffer_json_member_add_string(wb, "name", "TCP Outbound Connection");
- buffer_json_member_add_string(wb, "type", "line");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "connected_V4");
- buffer_json_add_array_item_string(wb, "connected_V6");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // TCP Functions
- buffer_json_member_add_object(wb, "TCPFunctions");
- {
- buffer_json_member_add_string(wb, "name", "TCPFunctions");
- buffer_json_member_add_string(wb, "type", "line");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "received");
- buffer_json_add_array_item_string(wb, "sent");
- buffer_json_add_array_item_string(wb, "close");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // TCP Bandwidth
- buffer_json_member_add_object(wb, "TCPBandwidth");
- {
- buffer_json_member_add_string(wb, "name", "TCPBandwidth");
- buffer_json_member_add_string(wb, "type", "line");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "received");
- buffer_json_add_array_item_string(wb, "sent");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // UDP Functions
- buffer_json_member_add_object(wb, "UDPFunctions");
- {
- buffer_json_member_add_string(wb, "name", "UDPFunctions");
- buffer_json_member_add_string(wb, "type", "line");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "received");
- buffer_json_add_array_item_string(wb, "sent");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // UDP Bandwidth
- buffer_json_member_add_object(wb, "UDPBandwidth");
- {
- buffer_json_member_add_string(wb, "name", "UDPBandwidth");
- buffer_json_member_add_string(wb, "type", "line");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "received");
- buffer_json_add_array_item_string(wb, "sent");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- }
- buffer_json_object_close(wb); // charts
-
- buffer_json_member_add_string(wb, "default_sort_column", "PID");
-
- // Do we use only on fields that can be groupped?
- buffer_json_member_add_object(wb, "group_by");
- {
- // group by PID
- buffer_json_member_add_object(wb, "PID");
- {
- buffer_json_member_add_string(wb, "name", "Process ID");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "PID");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // group by Process Name
- buffer_json_member_add_object(wb, "Process Name");
- {
- buffer_json_member_add_string(wb, "name", "Process Name");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Process Name");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // group by Process Name
- buffer_json_member_add_object(wb, "Origin");
- {
- buffer_json_member_add_string(wb, "name", "Origin");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Origin");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // group by Request From IP
- buffer_json_member_add_object(wb, "Request from");
- {
- buffer_json_member_add_string(wb, "name", "Request from IP");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Request from");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // group by Destination IP
- buffer_json_member_add_object(wb, "Destination IP");
- {
- buffer_json_member_add_string(wb, "name", "Destination IP");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Destination IP");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // group by DST Port
- buffer_json_member_add_object(wb, "Destination Port");
- {
- buffer_json_member_add_string(wb, "name", "Destination Port");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Destination Port");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
-
- // group by Protocol
- buffer_json_member_add_object(wb, "Protocol");
- {
- buffer_json_member_add_string(wb, "name", "Protocol");
- buffer_json_member_add_array(wb, "columns");
- {
- buffer_json_add_array_item_string(wb, "Protocol");
- }
- buffer_json_array_close(wb);
- }
- buffer_json_object_close(wb);
- }
- buffer_json_object_close(wb); // group_by
-
- buffer_json_member_add_time_t(wb, "expires", expires);
- buffer_json_finalize(wb);
-
- // Lock necessary to avoid race condition
- pluginsd_function_result_begin_to_stdout(transaction, HTTP_RESP_OK, "application/json", expires);
-
- fwrite(buffer_tostring(wb), buffer_strlen(wb), 1, stdout);
-
- pluginsd_function_result_end_to_stdout();
- fflush(stdout);
-
- buffer_free(wb);
-}
-
-/*****************************************************************
- * EBPF FUNCTION THREAD
- *****************************************************************/
-
-/**
- * FUNCTION thread.
- *
- * @param ptr a `ebpf_module_t *`.
- *
- * @return always NULL.
- */
-void *ebpf_function_thread(void *ptr)
-{
- (void)ptr;
-
- struct functions_evloop_globals *wg = functions_evloop_init(1,
- "EBPF",
- &lock,
- &ebpf_plugin_exit);
-
- functions_evloop_add_function(wg,
- "ebpf_socket",
- ebpf_function_socket_manipulation,
- PLUGINS_FUNCTIONS_TIMEOUT_DEFAULT);
-
- heartbeat_t hb;
- heartbeat_init(&hb);
- while(!ebpf_plugin_exit) {
- (void)heartbeat_next(&hb, USEC_PER_SEC);
-
- if (ebpf_plugin_exit) {
- break;
- }
- }
-
- return NULL;
-}
diff --git a/collectors/ebpf.plugin/ebpf_functions.h b/collectors/ebpf.plugin/ebpf_functions.h
deleted file mode 100644
index 795703b42..000000000
--- a/collectors/ebpf.plugin/ebpf_functions.h
+++ /dev/null
@@ -1,44 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-#ifndef NETDATA_EBPF_FUNCTIONS_H
-#define NETDATA_EBPF_FUNCTIONS_H 1
-
-#ifdef NETDATA_DEV_MODE
-// Common
-static inline void EBPF_PLUGIN_FUNCTIONS(const char *NAME, const char *DESC) {
- fprintf(stdout, "%s \"%s\" 10 \"%s\"\n", PLUGINSD_KEYWORD_FUNCTION, NAME, DESC);
-}
-#endif
-
-// configuration file & description
-#define NETDATA_DIRECTORY_FUNCTIONS_CONFIG_FILE "functions.conf"
-#define NETDATA_EBPF_FUNCTIONS_MODULE_DESC "Show information about current function status."
-
-// function list
-#define EBPF_FUNCTION_THREAD "ebpf_thread"
-#define EBPF_FUNCTION_SOCKET "ebpf_socket"
-
-// thread constants
-#define EBPF_PLUGIN_THREAD_FUNCTION_DESCRIPTION "Detailed information about eBPF threads."
-#define EBPF_PLUGIN_THREAD_FUNCTION_ERROR_THREAD_NOT_FOUND "ebpf.plugin does not have thread named "
-
-#define EBPF_THREADS_SELECT_THREAD "thread:"
-#define EBPF_THREADS_ENABLE_CATEGORY "enable:"
-#define EBPF_THREADS_DISABLE_CATEGORY "disable:"
-
-#define EBPF_THREAD_STATUS_RUNNING "running"
-#define EBPF_THREAD_STATUS_STOPPED "stopped"
-
-// socket constants
-#define EBPF_PLUGIN_SOCKET_FUNCTION_DESCRIPTION "Detailed information about open sockets."
-#define EBPF_FUNCTION_SOCKET_FAMILY "family:"
-#define EBPF_FUNCTION_SOCKET_PERIOD "period:"
-#define EBPF_FUNCTION_SOCKET_RESOLVE "resolve:"
-#define EBPF_FUNCTION_SOCKET_RANGE "range:"
-#define EBPF_FUNCTION_SOCKET_PORT "port:"
-#define EBPF_FUNCTION_SOCKET_RESET "reset"
-#define EBPF_FUNCTION_SOCKET_INTERFACES "interfaces"
-
-void *ebpf_function_thread(void *ptr);
-
-#endif
diff --git a/collectors/ebpf.plugin/metadata.yaml b/collectors/ebpf.plugin/metadata.yaml
deleted file mode 100644
index 97b5df389..000000000
--- a/collectors/ebpf.plugin/metadata.yaml
+++ /dev/null
@@ -1,3320 +0,0 @@
-plugin_name: ebpf.plugin
-modules:
- - meta:
- plugin_name: ebpf.plugin
- module_name: filedescriptor
- monitored_instance:
- name: eBPF Filedescriptor
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list:
- - plugin_name: apps.plugin
- module_name: apps
- - plugin_name: cgroups.plugin
- module_name: cgroups
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - file
- - eBPF
- - fd
- - open
- - close
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor calls for functions responsible to open or close a file descriptor and possible errors."
- method_description: "Attach tracing (kprobe and trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netdata sets necessary permissions during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "Depending of kernel version and frequency that files are open and close, this thread will add overhead every time that an internal kernel function monitored by this thread is called. The estimated additional period of time is between 90-200ms per call on kernels that do not have BTF technology."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- configuration:
- file:
- name: "ebpf.d/fd.conf"
- description: "Overwrite default configuration helping to reduce memory usage. You can also select charts visible on dashboard."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: apps
- description: Enable or disable integration with apps.plugin
- default_value: no
- required: false
- - name: cgroups
- description: Enable or disable integration with cgroup.plugin
- default_value: no
- required: false
- - name: pid table size
- description: Number of elements stored inside hash tables used to monitor calls per PID.
- default_value: 32768
- required: false
- - name: ebpf type format
- description: "Define the file type to load an eBPF program. Three options are available: `legacy` (Attach only `kprobe`), `co-re` (Plugin tries to use `trampoline` when available), and `auto` (plugin check OS configuration before to load)."
- default_value: auto
- required: false
- - name: ebpf co-re tracing
- description: "Select the attach method used by plugin when `co-re` is defined in previous option. Two options are available: `trampoline` (Option with lowest overhead), and `probe` (the same of legacy code)."
- default_value: trampoline
- required: false
- - name: maps per core
- description: Define how plugin will load their hash maps. When enabled (`yes`) plugin will load one hash table per core, instead to have centralized information.
- default_value: yes
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: cgroup
- description: "These Metrics show grouped information per cgroup/service."
- labels: []
- metrics:
- - name: cgroup.fd_open
- description: Number of open files
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: open
- - name: cgroup.fd_open_error
- description: Fails to open files
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: open
- - name: cgroup.fd_closed
- description: Files closed
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: close
- - name: cgroup.fd_close_error
- description: Fails to close files
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: close
- - name: services.file_open
- description: Number of open files
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.file_open_error
- description: Fails to open files
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.file_closed
- description: Files closed
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.file_close_error
- description: Fails to close files
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: global
- description: "These metrics show total number of calls to functions inside kernel."
- labels: []
- metrics:
- - name: filesystem.file_descriptor
- description: Open and close calls
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: open
- - name: close
- - name: filesystem.file_error
- description: Open fails
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: open
- - name: close
- - name: apps
- description: "These Metrics show grouped information per apps group."
- labels:
- - name: app_group
- description: The name of the group defined in the configuration.
- metrics:
- - name: app.ebpf_file_open
- description: Number of open files
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_file_open_error
- description: Fails to open files
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_file_closed
- description: Files closed
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_file_close_error
- description: Fails to close files
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - meta:
- plugin_name: ebpf.plugin
- module_name: processes
- monitored_instance:
- name: eBPF Processes
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list:
- - plugin_name: apps.plugin
- module_name: apps
- - plugin_name: cgroups.plugin
- module_name: cgroups
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - thread
- - fork
- - process
- - eBPF
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor calls for function creating tasks (threads and processes) inside Linux kernel."
- method_description: "Attach tracing (kprobe or tracepoint, and trampoline) to internal kernel functions."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT), files inside debugfs, and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- - title: Debug Filesystem
- description: |
- This thread needs to attach a tracepoint to monitor when a process schedule an exit event. To allow this specific feaure, it is necessary to mount `debugfs` (`mount -t debugfs none /sys/kernel/debug/`).
- configuration:
- file:
- name: "ebpf.d/process.conf"
- description: "Overwrite default configuration helping to reduce memory usage. You can also select charts visible on dashboard."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: apps
- description: Enable or disable integration with apps.plugin
- default_value: no
- required: false
- - name: cgroups
- description: Enable or disable integration with cgroup.plugin
- default_value: no
- required: false
- - name: pid table size
- description: Number of elements stored inside hash tables used to monitor calls per PID.
- default_value: 32768
- required: false
- - name: ebpf type format
- description: "Define the file type to load an eBPF program. Three options are available: `legacy` (Attach only `kprobe`), `co-re` (Plugin tries to use `trampoline` when available), and `auto` (plugin check OS configuration before to load)."
- default_value: auto
- required: false
- - name: ebpf co-re tracing
- description: "Select the attach method used by plugin when `co-re` is defined in previous option. Two options are available: `trampoline` (Option with lowest overhead), and `probe` (the same of legacy code). This plugin will always try to attach a tracepoint, so option here will impact only function used to monitor task (thread and process) creation."
- default_value: trampoline
- required: false
- - name: maps per core
- description: Define how plugin will load their hash maps. When enabled (`yes`) plugin will load one hash table per core, instead to have centralized information.
- default_value: yes
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show total number of calls to functions inside kernel."
- labels: []
- metrics:
- - name: system.process_thread
- description: Start process
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: process
- - name: system.process_status
- description: Process not closed
- unit: "difference"
- chart_type: line
- dimensions:
- - name: process
- - name: zombie
- - name: system.exit
- description: Exit process
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: process
- - name: system.task_error
- description: Fails to create process
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: task
- - name: apps
- description: "These Metrics show grouped information per apps group."
- labels:
- - name: app_group
- description: The name of the group defined in the configuration.
- metrics:
- - name: app.process_create
- description: Process started
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.thread_create
- description: Threads started
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: call
- - name: app.task_exit
- description: Tasks starts exit process
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: call
- - name: app.task_close
- description: Tasks closed
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: call
- - name: app.task_error
- description: Errors to create process or threads
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: app
- - name: cgroup
- description: "These Metrics show grouped information per cgroup/service."
- labels: []
- metrics:
- - name: cgroup.process_create
- description: Process started
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: process
- - name: cgroup.thread_create
- description: Threads started
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: thread
- - name: cgroup.task_exit
- description: Tasks starts exit process
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: exit
- - name: cgroup.task_close
- description: Tasks closed
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: process
- - name: cgroup.task_error
- description: Errors to create process or threads
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: process
- - name: services.process_create
- description: Process started
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.thread_create
- description: Threads started
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.task_close
- description: Tasks starts exit process
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.task_exit
- description: Tasks closed
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.task_error
- description: Errors to create process or threads
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - meta:
- plugin_name: ebpf.plugin
- module_name: disk
- monitored_instance:
- name: eBPF Disk
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - hard Disk
- - eBPF
- - latency
- - partition
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Measure latency for I/O events on disk."
- method_description: "Attach tracepoints to internal kernel functions."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT), files inside debugfs, and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- - title: Debug Filesystem
- description: |
- This thread needs to attach a tracepoint to monitor when a process schedule an exit event. To allow this specific feaure, it is necessary to mount `debugfs` (`mount -t debugfs none /sys/kernel/debug/`).`
- configuration:
- file:
- name: "ebpf.d/disk.conf"
- description: "Overwrite default configuration reducing number of I/O events."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: disk
- description: "These metrics measure latency for I/O events on every hard disk present on host."
- labels: []
- metrics:
- - name: disk.latency_io
- description: Disk latency
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: latency
- - meta:
- plugin_name: ebpf.plugin
- module_name: hardirq
- monitored_instance:
- name: eBPF Hardirq
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - HardIRQ
- - eBPF
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor latency for each HardIRQ available."
- method_description: "Attach tracepoints to internal kernel functions."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT), files inside debugfs, and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- - title: Debug Filesystem
- description: |
- This thread needs to attach a tracepoint to monitor when a process schedule an exit event. To allow this specific feaure, it is necessary to mount `debugfs` (`mount -t debugfs none /sys/kernel/debug/`).
- configuration:
- file:
- name: "ebpf.d/hardirq.conf"
- description: "Overwrite default configuration reducing number of I/O events."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show latest timestamp for each hardIRQ available on host."
- labels: []
- metrics:
- - name: system.hardirq_latency
- description: Hard IRQ latency
- unit: "milliseconds"
- chart_type: stacked
- dimensions:
- - name: hardirq names
- - meta:
- plugin_name: ebpf.plugin
- module_name: cachestat
- monitored_instance:
- name: eBPF Cachestat
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list:
- - plugin_name: apps.plugin
- module_name: apps
- - plugin_name: cgroups.plugin
- module_name: cgroups
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - Page cache
- - Hit ratio
- - eBPF
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor Linux page cache events giving for users a general vision about how his kernel is manipulating files."
- method_description: "Attach tracing (kprobe, trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called. The estimated additional period of time is between 90-200ms per call on kernels that do not have BTF technology."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- configuration:
- file:
- name: "ebpf.d/cachestat.conf"
- description: "Overwrite default configuration helping to reduce memory usage. You can also select charts visible on dashboard."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: apps
- description: Enable or disable integration with apps.plugin
- default_value: no
- required: false
- - name: cgroups
- description: Enable or disable integration with cgroup.plugin
- default_value: no
- required: false
- - name: pid table size
- description: Number of elements stored inside hash tables used to monitor calls per PID.
- default_value: 32768
- required: false
- - name: ebpf type format
- description: "Define the file type to load an eBPF program. Three options are available: `legacy` (Attach only `kprobe`), `co-re` (Plugin tries to use `trampoline` when available), and `auto` (plugin check OS configuration before to load)."
- default_value: auto
- required: false
- - name: ebpf co-re tracing
- description: "Select the attach method used by plugin when `co-re` is defined in previous option. Two options are available: `trampoline` (Option with lowest overhead), and `probe` (the same of legacy code)."
- default_value: trampoline
- required: false
- - name: maps per core
- description: Define how plugin will load their hash maps. When enabled (`yes`) plugin will load one hash table per core, instead to have centralized information.
- default_value: yes
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show total number of calls to functions inside kernel."
- labels: []
- metrics:
- - name: mem.cachestat_ratio
- description: Hit ratio
- unit: "%"
- chart_type: line
- dimensions:
- - name: ratio
- - name: mem.cachestat_dirties
- description: Number of dirty pages
- unit: "page/s"
- chart_type: line
- dimensions:
- - name: dirty
- - name: mem.cachestat_hits
- description: Number of accessed files
- unit: "hits/s"
- chart_type: line
- dimensions:
- - name: hit
- - name: mem.cachestat_misses
- description: Files out of page cache
- unit: "misses/s"
- chart_type: line
- dimensions:
- - name: miss
- - name: apps
- description: "These Metrics show grouped information per apps group."
- labels:
- - name: app_group
- description: The name of the group defined in the configuration.
- metrics:
- - name: app.ebpf_cachestat_hit_ratio
- description: Hit ratio
- unit: "%"
- chart_type: line
- dimensions:
- - name: ratio
- - name: app.ebpf_cachestat_dirty_pages
- description: Number of dirty pages
- unit: "page/s"
- chart_type: stacked
- dimensions:
- - name: pages
- - name: app.ebpf_cachestat_access
- description: Number of accessed files
- unit: "hits/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: app.ebpf_cachestat_misses
- description: Files out of page cache
- unit: "misses/s"
- chart_type: stacked
- dimensions:
- - name: misses
- - name: cgroup
- description: ""
- labels: []
- metrics:
- - name: cgroup.cachestat_ratio
- description: Hit ratio
- unit: "%"
- chart_type: line
- dimensions:
- - name: ratio
- - name: cgroup.cachestat_dirties
- description: Number of dirty pages
- unit: "page/s"
- chart_type: line
- dimensions:
- - name: dirty
- - name: cgroup.cachestat_hits
- description: Number of accessed files
- unit: "hits/s"
- chart_type: line
- dimensions:
- - name: hit
- - name: cgroup.cachestat_misses
- description: Files out of page cache
- unit: "misses/s"
- chart_type: line
- dimensions:
- - name: miss
- - name: services.cachestat_ratio
- description: Hit ratio
- unit: "%"
- chart_type: line
- dimensions:
- - name: a dimension per systemd service
- - name: services.cachestat_dirties
- description: Number of dirty pages
- unit: "page/s"
- chart_type: line
- dimensions:
- - name: a dimension per systemd service
- - name: services.cachestat_hits
- description: Number of accessed files
- unit: "hits/s"
- chart_type: line
- dimensions:
- - name: a dimension per systemd service
- - name: services.cachestat_misses
- description: Files out of page cache
- unit: "misses/s"
- chart_type: line
- dimensions:
- - name: a dimension per systemd service
- - meta:
- plugin_name: ebpf.plugin
- module_name: sync
- monitored_instance:
- name: eBPF Sync
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - syscall
- - eBPF
- - hard disk
- - memory
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor syscall responsible to move data from memory to storage device."
- method_description: "Attach tracing (kprobe, trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT, CONFIG_HAVE_SYSCALL_TRACEPOINTS), files inside debugfs, and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called. The estimated additional period of time is between 90-200ms per call on kernels that do not have BTF technology."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- - title: Debug Filesystem
- description: |
- This thread needs to attach a tracepoint to monitor when a process schedule an exit event. To allow this specific feaure, it is necessary to mount `debugfs` (`mount -t debugfs none /sys/kernel/debug`).
- configuration:
- file:
- name: "ebpf.d/sync.conf"
- description: "Overwrite default configuration and allows user to select charts visible on dashboard."
- options:
- description: |
- This configuration file have two different sections. The `[global]` overwrites all default options, while `[syscalls]` allow user to select the syscall to monitor.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: apps
- description: Enable or disable integration with apps.plugin
- default_value: no
- required: false
- - name: cgroups
- description: Enable or disable integration with cgroup.plugin
- default_value: no
- required: false
- - name: pid table size
- description: Number of elements stored inside hash tables used to monitor calls per PID.
- default_value: 32768
- required: false
- - name: ebpf type format
- description: "Define the file type to load an eBPF program. Three options are available: `legacy` (Attach only `kprobe`), `co-re` (Plugin tries to use `trampoline` when available), and `auto` (plugin check OS configuration before to load)."
- default_value: auto
- required: false
- - name: ebpf co-re tracing
- description: "Select the attach method used by plugin when `co-re` is defined in previous option. Two options are available: `trampoline` (Option with lowest overhead), and `probe` (the same of legacy code)."
- default_value: trampoline
- required: false
- - name: maps per core
- description: Define how plugin will load their hash maps. When enabled (`yes`) plugin will load one hash table per core, instead to have centralized information.
- default_value: yes
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- - name: sync
- description: Enable or disable monitoring for syscall `sync`
- default_value: yes
- required: false
- - name: msync
- description: Enable or disable monitoring for syscall `msync`
- default_value: yes
- required: false
- - name: fsync
- description: Enable or disable monitoring for syscall `fsync`
- default_value: yes
- required: false
- - name: fdatasync
- description: Enable or disable monitoring for syscall `fdatasync`
- default_value: yes
- required: false
- - name: syncfs
- description: Enable or disable monitoring for syscall `syncfs`
- default_value: yes
- required: false
- - name: sync_file_range
- description: Enable or disable monitoring for syscall `sync_file_range`
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: sync_freq
- link: https://github.com/netdata/netdata/blob/master/health/health.d/synchronization.conf
- metric: mem.sync
- info:
- number of sync() system calls. Every call causes all pending modifications to filesystem metadata and cached file data to be written to the
- underlying filesystems.
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show total number of calls to functions inside kernel."
- labels: []
- metrics:
- - name: mem.file_sync
- description: Monitor calls to fsync(2) and fdatasync(2).
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: fsync
- - name: fdatasync
- - name: mem.meory_map
- description: Monitor calls to msync(2).
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: msync
- - name: mem.sync
- description: Monitor calls to sync(2) and syncfs(2).
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: sync
- - name: syncfs
- - name: mem.file_segment
- description: Monitor calls to sync_file_range(2).
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: sync_file_range
- - meta:
- plugin_name: ebpf.plugin
- module_name: mdflush
- monitored_instance:
- name: eBPF MDflush
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - MD
- - RAID
- - eBPF
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor when flush events happen between disks."
- method_description: "Attach tracing (kprobe, trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that `md_flush_request` is called. The estimated additional period of time is between 90-200ms per call on kernels that do not have BTF technology."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- configuration:
- file:
- name: "ebpf.d/mdflush.conf"
- description: "Overwrite default configuration reducing I/O events."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "Number of times md_flush_request was called since last time."
- labels: []
- metrics:
- - name: mdstat.mdstat_flush
- description: MD flushes
- unit: "flushes"
- chart_type: stacked
- dimensions:
- - name: disk
- - meta:
- plugin_name: ebpf.plugin
- module_name: swap
- monitored_instance:
- name: eBPF SWAP
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list:
- - plugin_name: apps.plugin
- module_name: apps
- - plugin_name: cgroups.plugin
- module_name: cgroups
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - SWAP
- - memory
- - eBPF
- - Hard Disk
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitors when swap has I/O events and applications executing events."
- method_description: "Attach tracing (kprobe, trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called. The estimated additional period of time is between 90-200ms per call on kernels that do not have BTF technology."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- configuration:
- file:
- name: "ebpf.d/swap.conf"
- description: "Overwrite default configuration helping to reduce memory usage. You can also select charts visible on dashboard."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: apps
- description: Enable or disable integration with apps.plugin
- default_value: no
- required: false
- - name: cgroups
- description: Enable or disable integration with cgroup.plugin
- default_value: no
- required: false
- - name: pid table size
- description: Number of elements stored inside hash tables used to monitor calls per PID.
- default_value: 32768
- required: false
- - name: ebpf type format
- description: "Define the file type to load an eBPF program. Three options are available: `legacy` (Attach only `kprobe`), `co-re` (Plugin tries to use `trampoline` when available), and `auto` (plugin check OS configuration before to load)."
- default_value: auto
- required: false
- - name: ebpf co-re tracing
- description: "Select the attach method used by plugin when `co-re` is defined in previous option. Two options are available: `trampoline` (Option with lowest overhead), and `probe` (the same of legacy code)."
- default_value: trampoline
- required: false
- - name: maps per core
- description: Define how plugin will load their hash maps. When enabled (`yes`) plugin will load one hash table per core, instead to have centralized information.
- default_value: yes
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: cgroup
- description: "These Metrics show grouped information per cgroup/service."
- labels: []
- metrics:
- - name: cgroup.swap_read
- description: Calls to function swap_readpage.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: read
- - name: cgroup.swap_write
- description: Calls to function swap_writepage.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: write
- - name: services.swap_read
- description: Calls to swap_readpage.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.swap_write
- description: Calls to function swap_writepage.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: apps
- description: "These Metrics show grouped information per apps group."
- labels:
- - name: app_group
- description: The name of the group defined in the configuration.
- metrics:
- - name: app.ebpf_call_swap_readpage
- description: Calls to function swap_readpage.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per app group
- - name: app.ebpf_call_swap_writepage
- description: Calls to function swap_writepage.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per app group
- - name: global
- description: "These metrics show total number of calls to functions inside kernel."
- labels: []
- metrics:
- - name: mem.swapcalls
- description: Calls to access swap memory
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: write
- - name: read
- - meta:
- plugin_name: ebpf.plugin
- module_name: oomkill
- monitored_instance:
- name: eBPF OOMkill
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list:
- - plugin_name: apps.plugin
- module_name: apps
- - plugin_name: cgroups.plugin
- module_name: cgroups
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - application
- - memory
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor applications that reach out of memory."
- method_description: "Attach tracepoint to internal kernel functions."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT), files inside debugfs, and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- - title: Debug Filesystem
- description: |
- This thread needs to attach a tracepoint to monitor when a process schedule an exit event. To allow this specific feaure, it is necessary to mount `debugfs` (`mount -t debugfs none /sys/kernel/debug/`).
- configuration:
- file:
- name: "ebpf.d/oomkill.conf"
- description: "Overwrite default configuration reducing number of I/O events."
- options:
- description: |
- Overwrite default configuration reducing number of I/O events
- folding:
- title: "Config options"
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: cgroup
- description: "These metrics show cgroup/service that reached OOM."
- labels: []
- metrics:
- - name: cgroup.oomkills
- description: OOM kills. This chart is provided by eBPF plugin.
- unit: "kills"
- chart_type: line
- dimensions:
- - name: cgroup name
- - name: services.oomkills
- description: OOM kills. This chart is provided by eBPF plugin.
- unit: "kills"
- chart_type: line
- dimensions:
- - name: a dimension per systemd service
- - name: apps
- description: "These metrics show cgroup/service that reached OOM."
- labels:
- - name: app_group
- description: The name of the group defined in the configuration.
- metrics:
- - name: app.oomkill
- description: OOM kills
- unit: "kills"
- chart_type: stacked
- dimensions:
- - name: kills
- - meta:
- plugin_name: ebpf.plugin
- module_name: socket
- monitored_instance:
- name: eBPF Socket
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list:
- - plugin_name: apps.plugin
- module_name: apps
- - plugin_name: cgroups.plugin
- module_name: cgroups
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - TCP
- - UDP
- - bandwidth
- - server
- - connection
- - socket
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor bandwidth consumption per application for protocols TCP and UDP."
- method_description: "Attach tracing (kprobe, trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called. The estimated additional period of time is between 90-200ms per call on kernels that do not have BTF technology."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- configuration:
- file:
- name: "ebpf.d/network.conf"
- description: "Overwrite default configuration helping to reduce memory usage. You can also select charts visible on dashboard."
- options:
- description: |
- All options are defined inside section `[global]`. Options inside `network connections` are ignored for while.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: apps
- description: Enable or disable integration with apps.plugin
- default_value: no
- required: false
- - name: cgroups
- description: Enable or disable integration with cgroup.plugin
- default_value: no
- required: false
- - name: bandwidth table size
- description: Number of elements stored inside hash tables used to monitor calls per PID.
- default_value: 16384
- required: false
- - name: ipv4 connection table size
- description: Number of elements stored inside hash tables used to monitor calls per IPV4 connections.
- default_value: 16384
- required: false
- - name: ipv6 connection table size
- description: Number of elements stored inside hash tables used to monitor calls per IPV6 connections.
- default_value: 16384
- required: false
- - name: udp connection table size
- description: Number of temporary elements stored inside hash tables used to monitor UDP connections.
- default_value: 4096
- required: false
- - name: ebpf type format
- description: "Define the file type to load an eBPF program. Three options are available: `legacy` (Attach only `kprobe`), `co-re` (Plugin tries to use `trampoline` when available), and `auto` (plugin check OS configuration before to load)."
- default_value: auto
- required: false
- - name: ebpf co-re tracing
- description: "Select the attach method used by plugin when `co-re` is defined in previous option. Two options are available: `trampoline` (Option with lowest overhead), and `probe` (the same of legacy code)."
- default_value: trampoline
- required: false
- - name: maps per core
- description: Define how plugin will load their hash maps. When enabled (`yes`) plugin will load one hash table per core, instead to have centralized information.
- default_value: yes
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show total number of calls to functions inside kernel."
- labels: []
- metrics:
- - name: ip.inbound_conn
- description: Inbound connections.
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: connection_tcp
- - name: ip.tcp_outbound_conn
- description: TCP outbound connections.
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: received
- - name: ip.tcp_functions
- description: Calls to internal functions
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: received
- - name: send
- - name: closed
- - name: ip.total_tcp_bandwidth
- description: TCP bandwidth
- unit: "kilobits/s"
- chart_type: line
- dimensions:
- - name: received
- - name: send
- - name: ip.tcp_error
- description: TCP errors
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: received
- - name: send
- - name: ip.tcp_retransmit
- description: Packages retransmitted
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: retransmited
- - name: ip.udp_functions
- description: UDP calls
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: received
- - name: send
- - name: ip.total_udp_bandwidth
- description: UDP bandwidth
- unit: "kilobits/s"
- chart_type: line
- dimensions:
- - name: received
- - name: send
- - name: ip.udp_error
- description: UDP errors
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: received
- - name: send
- - name: apps
- description: "These metrics show grouped information per apps group."
- labels:
- - name: app_group
- description: The name of the group defined in the configuration.
- metrics:
- - name: app.ebpf_call_tcp_v4_connection
- description: Calls to tcp_v4_connection
- unit: "connections/s"
- chart_type: stacked
- dimensions:
- - name: connections
- - name: app.app.ebpf_call_tcp_v6_connection
- description: Calls to tcp_v6_connection
- unit: "connections/s"
- chart_type: stacked
- dimensions:
- - name: connections
- - name: app.ebpf_sock_bytes_sent
- description: Bytes sent
- unit: "kilobits/s"
- chart_type: stacked
- dimensions:
- - name: bandwidth
- - name: app.ebpf_sock_bytes_received
- description: bytes received
- unit: "kilobits/s"
- chart_type: stacked
- dimensions:
- - name: bandwidth
- - name: app.ebpf_call_tcp_sendmsg
- description: Calls for tcp_sendmsg
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_tcp_cleanup_rbuf
- description: Calls for tcp_cleanup_rbuf
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_tcp_retransmit
- description: Calls for tcp_retransmit
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_udp_sendmsg
- description: Calls for udp_sendmsg
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_udp_recvmsg
- description: Calls for udp_recvmsg
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: cgroup
- description: ""
- labels: []
- metrics:
- - name: cgroup.net_conn_ipv4
- description: Calls to tcp_v4_connection
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: connected_v4
- - name: cgroup.net_conn_ipv6
- description: Calls to tcp_v6_connection
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: connected_v6
- - name: cgroup.net_bytes_recv
- description: Bytes received
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: received
- - name: cgroup.net_bytes_sent
- description: Bytes sent
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: sent
- - name: cgroup.net_tcp_recv
- description: Calls to tcp_cleanup_rbuf.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: received
- - name: cgroup.net_tcp_send
- description: Calls to tcp_sendmsg.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: sent
- - name: cgroup.net_retransmit
- description: Calls to tcp_retransmit.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: retransmitted
- - name: cgroup.net_udp_send
- description: Calls to udp_sendmsg
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: sent
- - name: cgroup.net_udp_recv
- description: Calls to udp_recvmsg
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: received
- - name: services.net_conn_ipv6
- description: Calls to tcp_v6_connection
- unit: "connections/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.net_bytes_recv
- description: Bytes received
- unit: "kilobits/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.net_bytes_sent
- description: Bytes sent
- unit: "kilobits/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.net_tcp_recv
- description: Calls to tcp_cleanup_rbuf.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.net_tcp_send
- description: Calls to tcp_sendmsg.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.net_tcp_retransmit
- description: Calls to tcp_retransmit
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.net_udp_send
- description: Calls to udp_sendmsg
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.net_udp_recv
- description: Calls to udp_recvmsg
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - meta:
- plugin_name: ebpf.plugin
- module_name: dcstat
- monitored_instance:
- name: eBPF DCstat
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list:
- - plugin_name: apps.plugin
- module_name: apps
- - plugin_name: cgroups.plugin
- module_name: cgroups
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - Directory Cache
- - File system
- - eBPF
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor directory cache events per application given an overall vision about files on memory or storage device."
- method_description: "Attach tracing (kprobe, trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called. The estimated additional period of time is between 90-200ms per call on kernels that do not have BTF technology."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- configuration:
- file:
- name: "ebpf.d/dcstat.conf"
- description: "Overwrite default configuration helping to reduce memory usage. You can also select charts visible on dashboard."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config option"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: apps
- description: Enable or disable integration with apps.plugin
- default_value: no
- required: false
- - name: cgroups
- description: Enable or disable integration with cgroup.plugin
- default_value: no
- required: false
- - name: pid table size
- description: Number of elements stored inside hash tables used to monitor calls per PID.
- default_value: 32768
- required: false
- - name: ebpf type format
- description: "Define the file type to load an eBPF program. Three options are available: `legacy` (Attach only `kprobe`), `co-re` (Plugin tries to use `trampoline` when available), and `auto` (plugin check OS configuration before to load)."
- default_value: auto
- required: false
- - name: ebpf co-re tracing
- description: "Select the attach method used by plugin when `co-re` is defined in previous option. Two options are available: `trampoline` (Option with lowest overhead), and `probe` (the same of legacy code)."
- default_value: trampoline
- required: false
- - name: maps per core
- description: Define how plugin will load their hash maps. When enabled (`yes`) plugin will load one hash table per core, instead to have centralized information.
- default_value: yes
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: apps
- description: "These Metrics show grouped information per apps group."
- labels:
- - name: app_group
- description: The name of the group defined in the configuration.
- metrics:
- - name: app.ebpf_dc_ratio
- description: Percentage of files inside directory cache
- unit: "%"
- chart_type: line
- dimensions:
- - name: ratio
- - name: app.ebpf_dc_reference
- description: Count file access
- unit: "files"
- chart_type: stacked
- dimensions:
- - name: files
- - name: app.ebpf_dc_not_cache
- description: Files not present inside directory cache
- unit: "files"
- chart_type: stacked
- dimensions:
- - name: files
- - name: app.ebpf_dc_not_found
- description: Files not found
- unit: "files"
- chart_type: stacked
- dimensions:
- - name: files
- - name: filesystem
- description: "These metrics show total number of calls to functions inside kernel."
- labels: []
- metrics:
- - name: filesystem.dc_reference
- description: Variables used to calculate hit ratio.
- unit: "files"
- chart_type: line
- dimensions:
- - name: reference
- - name: slow
- - name: miss
- - name: filesystem.dc_hit_ratio
- description: Percentage of files inside directory cache
- unit: "%"
- chart_type: line
- dimensions:
- - name: ratio
- - name: cgroup
- description: ""
- labels: []
- metrics:
- - name: cgroup.dc_ratio
- description: Percentage of files inside directory cache
- unit: "%"
- chart_type: line
- dimensions:
- - name: ratio
- - name: cgroup.dc_reference
- description: Count file access
- unit: "files"
- chart_type: line
- dimensions:
- - name: reference
- - name: cgroup.dc_not_cache
- description: Files not present inside directory cache
- unit: "files"
- chart_type: line
- dimensions:
- - name: slow
- - name: cgroup.dc_not_found
- description: Files not found
- unit: "files"
- chart_type: line
- dimensions:
- - name: miss
- - name: services.dc_ratio
- description: Percentage of files inside directory cache
- unit: "%"
- chart_type: line
- dimensions:
- - name: a dimension per systemd service
- - name: services.dc_reference
- description: Count file access
- unit: "files"
- chart_type: line
- dimensions:
- - name: a dimension per systemd service
- - name: services.dc_not_cache
- description: Files not present inside directory cache
- unit: "files"
- chart_type: line
- dimensions:
- - name: a dimension per systemd service
- - name: services.dc_not_found
- description: Files not found
- unit: "files"
- chart_type: line
- dimensions:
- - name: a dimension per systemd service
- - meta:
- plugin_name: ebpf.plugin
- module_name: filesystem
- monitored_instance:
- name: eBPF Filesystem
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - Filesystem
- - ext4
- - btrfs
- - nfs
- - xfs
- - zfs
- - eBPF
- - latency
- - I/O
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor latency for main actions on filesystem like I/O events."
- method_description: "Attach tracing (kprobe, trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT), files inside debugfs, and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- configuration:
- file:
- name: "ebpf.d/filesystem.conf"
- description: "Overwrite default configuration and allows user to select charts visible on dashboard."
- options:
- description: |
- This configuration file have two different sections. The `[global]` overwrites default options, while `[filesystem]` allow user to select the filesystems to monitor.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- - name: btrfsdist
- description: Enable or disable latency monitoring for functions associated with btrfs filesystem.
- default_value: yes
- required: false
- - name: ext4dist
- description: Enable or disable latency monitoring for functions associated with ext4 filesystem.
- default_value: yes
- required: false
- - name: nfsdist
- description: Enable or disable latency monitoring for functions associated with nfs filesystem.
- default_value: yes
- required: false
- - name: xfsdist
- description: Enable or disable latency monitoring for functions associated with xfs filesystem.
- default_value: yes
- required: false
- - name: zfsdist
- description: Enable or disable latency monitoring for functions associated with zfs filesystem.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: filesystem
- description: "Latency charts associate with filesystem actions."
- labels: []
- metrics:
- - name: filesystem.read_latency
- description: ext4 latency for each read request.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: latency period
- - name: filesystem.open_latency
- description: ext4 latency for each open request.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: latency period
- - name: filesystem.sync_latency
- description: ext4 latency for each sync request.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: latency period
- - name: iilesystem
- description: ""
- labels: []
- metrics:
- - name: filesystem.write_latency
- description: ext4 latency for each write request.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: latency period
- - name: global
- description: ""
- labels: []
- metrics:
- - name: filesystem.attributte_latency
- description: nfs latency for each attribute request.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: latency period
- - meta:
- plugin_name: ebpf.plugin
- module_name: shm
- monitored_instance:
- name: eBPF SHM
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list:
- - plugin_name: apps.plugin
- module_name: apps
- - plugin_name: cgroups.plugin
- module_name: cgroups
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - syscall
- - shared memory
- - eBPF
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor syscall responsible to manipulate shared memory."
- method_description: "Attach tracing (kprobe, trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called. The estimated additional period of time is between 90-200ms per call on kernels that do not have BTF technology."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- - title: Debug Filesystem
- description: |
- This thread needs to attach a tracepoint to monitor when a process schedule an exit event. To allow this specific feaure, it is necessary to mount `debugfs` (`mount -t debugfs none /sys/kernel/debug/`).`
- configuration:
- file:
- name: "ebpf.d/shm.conf"
- description: "Overwrite default configuration and allows user to select charts visible on dashboard."
- options:
- description: |
- This configuration file have two different sections. The `[global]` overwrites all default options, while `[syscalls]` allow user to select the syscall to monitor.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: apps
- description: Enable or disable integration with apps.plugin
- default_value: no
- required: false
- - name: cgroups
- description: Enable or disable integration with cgroup.plugin
- default_value: no
- required: false
- - name: pid table size
- description: Number of elements stored inside hash tables used to monitor calls per PID.
- default_value: 32768
- required: false
- - name: ebpf type format
- description: "Define the file type to load an eBPF program. Three options are available: `legacy` (Attach only `kprobe`), `co-re` (Plugin tries to use `trampoline` when available), and `auto` (plugin check OS configuration before to load)."
- default_value: auto
- required: false
- - name: ebpf co-re tracing
- description: "Select the attach method used by plugin when `co-re` is defined in previous option. Two options are available: `trampoline` (Option with lowest overhead), and `probe` (the same of legacy code)."
- default_value: trampoline
- required: false
- - name: maps per core
- description: Define how plugin will load their hash maps. When enabled (`yes`) plugin will load one hash table per core, instead to have centralized information.
- default_value: yes
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- - name: shmget
- description: Enable or disable monitoring for syscall `shmget`
- default_value: yes
- required: false
- - name: shmat
- description: Enable or disable monitoring for syscall `shmat`
- default_value: yes
- required: false
- - name: shmdt
- description: Enable or disable monitoring for syscall `shmdt`
- default_value: yes
- required: false
- - name: shmctl
- description: Enable or disable monitoring for syscall `shmctl`
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: cgroup
- description: "These Metrics show grouped information per cgroup/service."
- labels: []
- metrics:
- - name: cgroup.shmget
- description: Calls to syscall shmget(2).
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: get
- - name: cgroup.shmat
- description: Calls to syscall shmat(2).
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: at
- - name: cgroup.shmdt
- description: Calls to syscall shmdt(2).
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: dt
- - name: cgroup.shmctl
- description: Calls to syscall shmctl(2).
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: ctl
- - name: services.shmget
- description: Calls to syscall shmget(2).
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.shmat
- description: Calls to syscall shmat(2).
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.shmdt
- description: Calls to syscall shmdt(2).
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.shmctl
- description: Calls to syscall shmctl(2).
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: apps
- description: "These Metrics show grouped information per apps group."
- labels:
- - name: app_group
- description: The name of the group defined in the configuration.
- metrics:
- - name: app.ebpf_shmget_call
- description: Calls to syscall shmget(2).
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_shmat_call
- description: Calls to syscall shmat(2).
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_shmdt_call
- description: Calls to syscall shmdt(2).
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_shmctl_call
- description: Calls to syscall shmctl(2).
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: global
- description: "These Metrics show number of calls for specified syscall."
- labels: []
- metrics:
- - name: system.shared_memory_calls
- description: Calls to shared memory system calls
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: get
- - name: at
- - name: dt
- - name: ctl
- - meta:
- plugin_name: ebpf.plugin
- module_name: softirq
- monitored_instance:
- name: eBPF SoftIRQ
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - SoftIRQ
- - eBPF
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor latency for each SoftIRQ available."
- method_description: "Attach kprobe to internal kernel functions."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT), files inside debugfs, and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- - title: Debug Filesystem
- description: |
- This thread needs to attach a tracepoint to monitor when a process schedule an exit event. To allow this specific feaure, it is necessary to mount `debugfs` (`mount -t debugfs none /sys/kernel/debug/`).`
- configuration:
- file:
- name: "ebpf.d/softirq.conf"
- description: "Overwrite default configuration reducing number of I/O events."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show latest timestamp for each softIRQ available on host."
- labels: []
- metrics:
- - name: system.softirq_latency
- description: Soft IRQ latency
- unit: "milliseconds"
- chart_type: stacked
- dimensions:
- - name: soft IRQs
- - meta:
- plugin_name: ebpf.plugin
- module_name: mount
- monitored_instance:
- name: eBPF Mount
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - mount
- - umount
- - device
- - eBPF
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor calls for mount and umount syscall."
- method_description: "Attach tracing (kprobe, trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT, CONFIG_HAVE_SYSCALL_TRACEPOINTS), files inside debugfs, and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called. The estimated additional period of time is between 90-200ms per call on kernels that do not have BTF technology."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- - title: Debug Filesystem
- description: |
- This thread needs to attach a tracepoint to monitor when a process schedule an exit event. To allow this specific feaure, it is necessary to mount `debugfs` (`mount -t debugfs none /sys/kernel/debug/`).`
- configuration:
- file:
- name: "ebpf.d/mount.conf"
- description: "Overwrite default configuration."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: ebpf type format
- description: "Define the file type to load an eBPF program. Three options are available: `legacy` (Attach only `kprobe`), `co-re` (Plugin tries to use `trampoline` when available), and `auto` (plugin check OS configuration before to load)."
- default_value: auto
- required: false
- - name: ebpf co-re tracing
- description: "Select the attach method used by plugin when `co-re` is defined in previous option. Two options are available: `trampoline` (Option with lowest overhead), and `probe` (the same of legacy code)."
- default_value: trampoline
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "Calls for syscalls mount an umount."
- labels: []
- metrics:
- - name: mount_points.call
- description: Calls to mount and umount syscalls
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: mount
- - name: umount
- - name: mount_points.error
- description: Errors to mount and umount file systems
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: mount
- - name: umount
- - meta:
- plugin_name: ebpf.plugin
- module_name: vfs
- monitored_instance:
- name: eBPF VFS
- link: "https://kernel.org/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list:
- - plugin_name: apps.plugin
- module_name: apps
- - plugin_name: cgroups.plugin
- module_name: cgroups
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - virtual
- - filesystem
- - eBPF
- - I/O
- - files
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor I/O events on Linux Virtual Filesystem."
- method_description: "Attach tracing (kprobe, trampoline) to internal kernel functions according options used to compile kernel."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid because it loads data inside kernel. Netada sets necessary permission during installation time."
- default_behavior:
- auto_detection:
- description: "The plugin checks kernel compilation flags (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) and presence of BTF files to decide which eBPF program will be attached."
- limits:
- description: ""
- performance_impact:
- description: "This thread will add overhead every time that an internal kernel function monitored by this thread is called. The estimated additional period of time is between 90-200ms per call on kernels that do not have BTF technology."
- setup:
- prerequisites:
- list:
- - title: Compile kernel
- description: |
- Check if your kernel was compiled with necessary options (CONFIG_KPROBES, CONFIG_BPF, CONFIG_BPF_SYSCALL, CONFIG_BPF_JIT) in `/proc/config.gz` or inside /boot/config file. Some cited names can be different accoring preferences of Linux distributions.
- When you do not have options set, it is necessary to get the kernel source code from https://kernel.org or a kernel package from your distribution, this last is preferred. The kernel compilation has a well definedd pattern, but distributions can deliver their configuration files
- with different names.
-
- Now follow steps:
- 1. Copy the configuration file to /usr/src/linux/.config.
- 2. Select the necessary options: make oldconfig
- 3. Compile your kernel image: make bzImage
- 4. Compile your modules: make modules
- 5. Copy your new kernel image for boot loader directory
- 6. Install the new modules: make modules_install
- 7. Generate an initial ramdisk image (`initrd`) if it is necessary.
- 8. Update your boot loader
- configuration:
- file:
- name: "ebpf.d/vfs.conf"
- description: "Overwrite default configuration helping to reduce memory usage."
- options:
- description: |
- All options are defined inside section `[global]`.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 5
- required: false
- - name: ebpf load mode
- description: Define whether plugin will monitor the call (`entry`) for the functions or it will also monitor the return (`return`).
- default_value: entry
- required: false
- - name: apps
- description: Enable or disable integration with apps.plugin
- default_value: no
- required: false
- - name: cgroups
- description: Enable or disable integration with cgroup.plugin
- default_value: no
- required: false
- - name: pid table size
- description: Number of elements stored inside hash tables used to monitor calls per PID.
- default_value: 32768
- required: false
- - name: ebpf type format
- description: "Define the file type to load an eBPF program. Three options are available: `legacy` (Attach only `kprobe`), `co-re` (Plugin tries to use `trampoline` when available), and `auto` (plugin check OS configuration before to load)."
- default_value: auto
- required: false
- - name: ebpf co-re tracing
- description: "Select the attach method used by plugin when `co-re` is defined in previous option. Two options are available: `trampoline` (Option with lowest overhead), and `probe` (the same of legacy code)."
- default_value: trampoline
- required: false
- - name: maps per core
- description: Define how plugin will load their hash maps. When enabled (`yes`) plugin will load one hash table per core, instead to have centralized information.
- default_value: yes
- required: false
- - name: lifetime
- description: Set default lifetime for thread when enabled by cloud.
- default_value: 300
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: cgroup
- description: "These Metrics show grouped information per cgroup/service."
- labels: []
- metrics:
- - name: cgroup.vfs_unlink
- description: Files deleted
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: delete
- - name: cgroup.vfs_write
- description: Write to disk
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: write
- - name: cgroup.vfs_write_error
- description: Fails to write
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: write
- - name: cgroup.vfs_read
- description: Read from disk
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: read
- - name: cgroup.vfs_read_error
- description: Fails to read
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: read
- - name: cgroup.vfs_write_bytes
- description: Bytes written on disk
- unit: "bytes/s"
- chart_type: line
- dimensions:
- - name: write
- - name: cgroup.vfs_read_bytes
- description: Bytes read from disk
- unit: "bytes/s"
- chart_type: line
- dimensions:
- - name: read
- - name: cgroup.vfs_fsync
- description: Calls to vfs_fsync.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: fsync
- - name: cgroup.vfs_fsync_error
- description: Sync error
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: fsync
- - name: cgroup.vfs_open
- description: Calls to vfs_open.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: open
- - name: cgroup.vfs_open_error
- description: Open error
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: open
- - name: cgroup.vfs_create
- description: Calls to vfs_create.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: create
- - name: cgroup.vfs_create_error
- description: Create error
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: create
- - name: services.vfs_unlink
- description: Files deleted
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_write
- description: Write to disk
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_write_error
- description: Fails to write
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_read
- description: Read from disk
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_read_error
- description: Fails to read
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_write_bytes
- description: Bytes written on disk
- unit: "bytes/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_read_bytes
- description: Bytes read from disk
- unit: "bytes/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_fsync
- description: Calls to vfs_fsync.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_fsync_error
- description: Sync error
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_open
- description: Calls to vfs_open.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_open_error
- description: Open error
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_create
- description: Calls to vfs_create.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: services.vfs_create_error
- description: Create error
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per systemd service
- - name: global
- description: "These Metrics show grouped information per cgroup/service."
- labels: []
- metrics:
- - name: filesystem.vfs_deleted_objects
- description: Remove files
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: delete
- - name: filesystem.vfs_io
- description: Calls to IO
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: filesystem.vfs_io_bytes
- description: Bytes written and read
- unit: "bytes/s"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: filesystem.vfs_io_error
- description: Fails to write or read
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: read
- - name: write
- - name: filesystem.vfs_fsync
- description: Calls to vfs_fsync.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: fsync
- - name: filesystem.vfs_fsync_error
- description: Fails to synchronize
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: fsync
- - name: filesystem.vfs_open
- description: Calls to vfs_open.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: open
- - name: filesystem.vfs_open_error
- description: Fails to open a file
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: open
- - name: filesystem.vfs_create
- description: Calls to vfs_create.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: create
- - name: filesystem.vfs_create_error
- description: Fails to create a file.
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: create
- - name: apps
- description: "These Metrics show grouped information per apps group."
- labels:
- - name: app_group
- description: The name of the group defined in the configuration.
- metrics:
- - name: app.ebpf_call_vfs_unlink
- description: Files deleted
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_vfs_write
- description: Write to disk
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_vfs_write_error
- description: Fails to write
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_vfs_read
- description: Read from disk
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_vfs_read_error
- description: Fails to read
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_vfs_write_bytes
- description: Bytes written on disk
- unit: "bytes/s"
- chart_type: stacked
- dimensions:
- - name: writes
- - name: app.ebpf_call_vfs_read_bytes
- description: Bytes read on disk
- unit: "bytes/s"
- chart_type: stacked
- dimensions:
- - name: reads
- - name: app.ebpf_call_vfs_fsync
- description: Calls to vfs_fsync.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_vfs_fsync_error
- description: Sync error
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_vfs_open
- description: Calls to vfs_open.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_vfs_open_error
- description: Open error
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_vfs_create
- description: Calls to vfs_create.
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - name: app.ebpf_call_vfs_create_error
- description: Create error
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: calls
- - meta:
- plugin_name: ebpf.plugin
- module_name: process
- monitored_instance:
- name: eBPF Process
- link: "https://github.com/netdata/netdata/"
- categories:
- - data-collection.ebpf
- icon_filename: "ebpf.jpg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - Memory
- - plugin
- - eBPF
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor internal memory usage."
- method_description: "Uses netdata internal statistic to monitor memory management by plugin."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: Netdata flags.
- description: "To have these charts you need to compile netdata with flag `NETDATA_DEV_MODE`."
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "How plugin is allocating memory."
- labels: []
- metrics:
- - name: netdata.ebpf_aral_stat_size
- description: Bytes allocated for ARAL.
- unit: "bytes"
- chart_type: stacked
- dimensions:
- - name: memory
- - name: netdata.ebpf_aral_stat_alloc
- description: Calls to allocate memory.
- unit: "calls"
- chart_type: stacked
- dimensions:
- - name: aral
- - name: netdata.ebpf_threads
- description: Threads info
- unit: "threads"
- chart_type: line
- dimensions:
- - name: total
- - name: running
- - name: netdata.ebpf_load_methods
- description: Load info
- unit: "methods"
- chart_type: line
- dimensions:
- - name: legacy
- - name: co-re
- - name: netdata.ebpf_kernel_memory
- description: Memory allocated for hash tables.
- unit: "bytes"
- chart_type: line
- dimensions:
- - name: memory_locked
- - name: netdata.ebpf_hash_tables_count
- description: Number of hash tables loaded
- unit: "hash tables"
- chart_type: line
- dimensions:
- - name: hash_table
- - name: netdata.ebpf_aral_stat_size
- description: Bytes allocated for ARAL
- unit: "bytes"
- chart_type: stacked
- dimensions:
- - name: memory
- - name: netdata.ebpf_aral_stat_alloc
- description: Calls to allocate memory
- unit: "calls"
- chart_type: stacked
- dimensions:
- - name: aral
- - name: netdata.ebpf_aral_stat_size
- description: Bytes allocated for ARAL.
- unit: "bytes"
- chart_type: stacked
- dimensions:
- - name: memory
- - name: netdata.ebpf_aral_stat_alloc
- description: Calls to allocate memory
- unit: "calls"
- chart_type: stacked
- dimensions:
- - name: aral
- - name: netdata.ebpf_hash_tables_insert_pid_elements
- description: Number of times an element was inserted in a hash table.
- unit: "rows"
- chart_type: line
- dimensions:
- - name: thread
- - name: netdata.ebpf_hash_tables_remove_pid_elements
- description: Number of times an element was removed in a hash table.
- unit: "rows"
- chart_type: line
- dimensions:
- - name: thread
diff --git a/collectors/freebsd.plugin/Makefile.am b/collectors/freebsd.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/freebsd.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/freebsd.plugin/README.md b/collectors/freebsd.plugin/README.md
deleted file mode 100644
index 9c33fccb1..000000000
--- a/collectors/freebsd.plugin/README.md
+++ /dev/null
@@ -1,16 +0,0 @@
-<!--
-title: "FreeBSD system metrics (freebsd.plugin)"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/README.md"
-sidebar_label: "FreeBSD system metrics (freebsd.plugin)"
-learn_status: "Published"
-learn_topic_type: "References"
-learn_rel_path: "Integrations/Monitor/System metrics"
--->
-
-# FreeBSD system metrics (freebsd.plugin)
-
-Collects resource usage and performance data on FreeBSD systems
-
-By default, Netdata will enable monitoring metrics for disks, memory, and network only when they are not zero. If they are constantly zero they are ignored. Metrics that will start having values, after Netdata is started, will be detected and charts will be automatically added to the dashboard (a refresh of the dashboard is needed for them to appear though). Use `yes` instead of `auto` in plugin configuration sections to enable these charts permanently. You can also set the `enable zero metrics` option to `yes` in the `[global]` section which enables charts with zero metrics for all internal Netdata plugins.
-
-
diff --git a/collectors/freebsd.plugin/metadata.yaml b/collectors/freebsd.plugin/metadata.yaml
deleted file mode 100644
index 36fba2430..000000000
--- a/collectors/freebsd.plugin/metadata.yaml
+++ /dev/null
@@ -1,3398 +0,0 @@
-plugin_name: freebsd.plugin
-modules:
- - meta:
- plugin_name: freebsd.plugin
- module_name: vm.loadavg
- monitored_instance:
- name: vm.loadavg
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "System Load Average"
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: vm.loadavg
- description: Enable or disable load average metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: load_cpu_number
- link: https://github.com/netdata/netdata/blob/master/health/health.d/load.conf
- metric: system.load
- info: number of active CPU cores in the system
- os: "linux"
- - name: load_average_15
- link: https://github.com/netdata/netdata/blob/master/health/health.d/load.conf
- metric: system.load
- info: system fifteen-minute load average
- os: "linux"
- - name: load_average_5
- link: https://github.com/netdata/netdata/blob/master/health/health.d/load.conf
- metric: system.load
- info: system five-minute load average
- os: "linux"
- - name: load_average_1
- link: https://github.com/netdata/netdata/blob/master/health/health.d/load.conf
- metric: system.load
- info: system one-minute load average
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "Monitoring for number of threads running or waiting."
- labels: []
- metrics:
- - name: system.load
- description: System Load Average
- unit: "load"
- chart_type: line
- dimensions:
- - name: load1
- - name: load5
- - name: load15
- - meta:
- plugin_name: freebsd.plugin
- module_name: vm.vmtotal
- monitored_instance:
- name: vm.vmtotal
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "memory.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect Virtual Memory information from host."
- method_description: "The plugin calls function `sysctl` to collect data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:vm.vmtotal]"
- description: ""
- options:
- description: ""
- folding:
- title: "Config Options"
- enabled: true
- list:
- - name: enable total processes
- description: Number of active processes.
- default_value: yes
- required: false
- - name: processes running
- description: Show number of processes running or blocked.
- default_value: yes
- required: false
- - name: real memory
- description: Memeory used on host.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: active_processes
- link: https://github.com/netdata/netdata/blob/master/health/health.d/processes.conf
- metric: system.active_processes
- info: system process IDs (PID) space utilization
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show an overall vision about processes running."
- labels: []
- metrics:
- - name: system.active_processes
- description: System Active Processes
- unit: "processes"
- chart_type: line
- dimensions:
- - name: active
- - name: system.processes
- description: System Processes
- unit: "processes"
- chart_type: line
- dimensions:
- - name: running
- - name: blocked
- - name: mem.real
- description: Total Real Memory In Use
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: used
- - meta:
- plugin_name: freebsd.plugin
- module_name: kern.cp_time
- monitored_instance:
- name: kern.cp_time
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Total CPU utilization"
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- description: "[plugin:freebsd]"
- options:
- description: "The netdata main configuration file."
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: kern.cp_time
- description: Enable or disable Total CPU usage.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 10min_cpu_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf
- metric: system.cpu
- info: average CPU utilization over the last 10 minutes (excluding iowait, nice and steal)
- os: "linux"
- - name: 10min_cpu_iowait
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf
- metric: system.cpu
- info: average CPU iowait time over the last 10 minutes
- os: "linux"
- - name: 20min_steal_cpu
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf
- metric: system.cpu
- info: average CPU steal time over the last 20 minutes
- os: "linux"
- - name: 10min_cpu_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf
- metric: system.cpu
- info: average CPU utilization over the last 10 minutes (excluding nice)
- os: "freebsd"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show CPU usage statistics."
- labels: []
- metrics:
- - name: system.cpu
- description: Total CPU utilization
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: nice
- - name: system
- - name: user
- - name: interrupt
- - name: idle
- - name: core
- description: ""
- labels: []
- metrics:
- - name: cpu.cpu
- description: Core utilization
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: nice
- - name: system
- - name: user
- - name: interrupt
- - name: idle
- - meta:
- plugin_name: freebsd.plugin
- module_name: dev.cpu.temperature
- monitored_instance:
- name: dev.cpu.temperature
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Get current CPU temperature"
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: dev.cpu.temperature
- description: Enable or disable CPU temperature metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "This metric show latest CPU temperature."
- labels: []
- metrics:
- - name: cpu.temperature
- description: Core temperature
- unit: "Celsius"
- chart_type: line
- dimensions:
- - name: a dimension per core
- - meta:
- plugin_name: freebsd.plugin
- module_name: dev.cpu.0.freq
- monitored_instance:
- name: dev.cpu.0.freq
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Read current CPU Scaling frequency."
- method_description: "Current CPU Scaling Frequency"
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "Config options"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file"
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list:
- - name: dev.cpu.0.freq
- description: Enable or disable CPU Scaling frequency metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "The metric shows status of CPU frequency, it is direct affected by system load."
- labels: []
- metrics:
- - name: cpu.scaling_cur_freq
- description: Current CPU Scaling Frequency
- unit: "MHz"
- chart_type: line
- dimensions:
- - name: frequency
- - meta:
- plugin_name: freebsd.plugin
- module_name: hw.intrcnt
- monitored_instance:
- name: hw.intrcnt
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Get total number of interrupts"
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config option"
- enabled: true
- list:
- - name: hw.intrcnt
- description: Enable or disable Interrupts metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show system interrupts frequency."
- labels: []
- metrics:
- - name: system.intr
- description: Total Hardware Interrupts
- unit: "interrupts/s"
- chart_type: line
- dimensions:
- - name: interrupts
- - name: system.interrupts
- description: System interrupts
- unit: "interrupts/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per interrupt
- - meta:
- plugin_name: freebsd.plugin
- module_name: vm.stats.sys.v_intr
- monitored_instance:
- name: vm.stats.sys.v_intr
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Device interrupts"
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config option"
- enabled: true
- list:
- - name: vm.stats.sys.v_intr
- description: Enable or disable device interrupts metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "The metric show device interrupt frequency."
- labels: []
- metrics:
- - name: system.dev_intr
- description: Device Interrupts
- unit: "interrupts/s"
- chart_type: line
- dimensions:
- - name: interrupts
- - meta:
- plugin_name: freebsd.plugin
- module_name: vm.stats.sys.v_soft
- monitored_instance:
- name: vm.stats.sys.v_soft
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Software Interrupt"
- method_description: "vm.stats.sys.v_soft"
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config option"
- enabled: true
- list:
- - name: vm.stats.sys.v_soft
- description: Enable or disable software inerrupts metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "This metric shows software interrupt frequency."
- labels: []
- metrics:
- - name: system.soft_intr
- description: Software Interrupts
- unit: "interrupts/s"
- chart_type: line
- dimensions:
- - name: interrupts
- - meta:
- plugin_name: freebsd.plugin
- module_name: vm.stats.sys.v_swtch
- monitored_instance:
- name: vm.stats.sys.v_swtch
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "CPU context switch"
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: vm.stats.sys.v_swtch
- description: Enable or disable CPU context switch metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "The metric count the number of context switches happening on host."
- labels: []
- metrics:
- - name: system.ctxt
- description: CPU Context Switches
- unit: "context switches/s"
- chart_type: line
- dimensions:
- - name: switches
- - name: system.forks
- description: Started Processes
- unit: "processes/s"
- chart_type: line
- dimensions:
- - name: started
- - meta:
- plugin_name: freebsd.plugin
- module_name: vm.swap_info
- monitored_instance:
- name: vm.swap_info
- link: ""
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect information about SWAP memory."
- method_description: "The plugin calls `sysctlnametomib` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: vm.swap_info
- description: Enable or disable SWAP metrics.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: used_swap
- link: https://github.com/netdata/netdata/blob/master/health/health.d/swap.conf
- metric: mem.swap
- info: swap memory utilization
- os: "linux freebsd"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "This metric shows the SWAP usage."
- labels: []
- metrics:
- - name: mem.swap
- description: System Swap
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: free
- - name: used
- - meta:
- plugin_name: freebsd.plugin
- module_name: system.ram
- monitored_instance:
- name: system.ram
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "memory.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Show information about system memory usage."
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: system.ram
- description: Enable or disable system RAM metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: ram_in_use
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf
- metric: system.ram
- info: system memory utilization
- os: "linux"
- - name: ram_in_use
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf
- metric: system.ram
- info: system memory utilization
- os: "freebsd"
- - name: ram_available
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf
- metric: mem.available
- info: percentage of estimated amount of RAM available for userspace processes, without causing swapping
- os: "linux"
- - name: ram_available
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf
- metric: mem.available
- info: percentage of estimated amount of RAM available for userspace processes, without causing swapping
- os: "freebsd"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "This metric shows RAM usage statistics."
- labels: []
- metrics:
- - name: system.ram
- description: System RAM
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: free
- - name: active
- - name: inactive
- - name: wired
- - name: cache
- - name: laundry
- - name: buffers
- - name: mem.available
- description: Available RAM for applications
- unit: "MiB"
- chart_type: line
- dimensions:
- - name: avail
- - meta:
- plugin_name: freebsd.plugin
- module_name: vm.stats.vm.v_swappgs
- monitored_instance:
- name: vm.stats.vm.v_swappgs
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "memory.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "The metric swap amount of data read from and written to SWAP."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: vm.stats.vm.v_swappgs
- description: Enable or disable infoormation about SWAP I/O metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 30min_ram_swapped_out
- link: https://github.com/netdata/netdata/blob/master/health/health.d/swap.conf
- metric: mem.swapio
- info: percentage of the system RAM swapped in the last 30 minutes
- os: "linux freebsd"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "This metric shows events happening on SWAP."
- labels: []
- metrics:
- - name: mem.swapio
- description: Swap I/O
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: io
- - name: out
- - meta:
- plugin_name: freebsd.plugin
- module_name: vm.stats.vm.v_pgfaults
- monitored_instance:
- name: vm.stats.vm.v_pgfaults
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "memory.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect memory page faults events."
- method_description: "The plugin calls `sysctl` function to collect necessary data"
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: vm.stats.vm.v_pgfaults
- description: Enable or disable Memory page fault metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "The number of page faults happened on host."
- labels: []
- metrics:
- - name: mem.pgfaults
- description: Memory Page Faults
- unit: "page faults/s"
- chart_type: line
- dimensions:
- - name: memory
- - name: io_requiring
- - name: cow
- - name: cow_optimized
- - name: in_transit
- - meta:
- plugin_name: freebsd.plugin
- module_name: kern.ipc.sem
- monitored_instance:
- name: kern.ipc.sem
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect information about semaphore."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: kern.ipc.sem
- description: Enable or disable semaphore metrics.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: semaphores_used
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ipc.conf
- metric: system.ipc_semaphores
- info: IPC semaphore utilization
- os: "linux"
- - name: semaphore_arrays_used
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ipc.conf
- metric: system.ipc_semaphore_arrays
- info: IPC semaphore arrays utilization
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics shows counters for semaphores on host."
- labels: []
- metrics:
- - name: system.ipc_semaphores
- description: IPC Semaphores
- unit: "semaphores"
- chart_type: area
- dimensions:
- - name: semaphores
- - name: system.ipc_semaphore_arrays
- description: IPC Semaphore Arrays
- unit: "arrays"
- chart_type: area
- dimensions:
- - name: arrays
- - meta:
- plugin_name: freebsd.plugin
- module_name: kern.ipc.shm
- monitored_instance:
- name: kern.ipc.shm
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "memory.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect shared memory information."
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: kern.ipc.shm
- description: Enable or disable shared memory metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics give status about current shared memory segments."
- labels: []
- metrics:
- - name: system.ipc_shared_mem_segs
- description: IPC Shared Memory Segments
- unit: "segments"
- chart_type: area
- dimensions:
- - name: segments
- - name: system.ipc_shared_mem_size
- description: IPC Shared Memory Segments Size
- unit: "KiB"
- chart_type: area
- dimensions:
- - name: allocated
- - meta:
- plugin_name: freebsd.plugin
- module_name: kern.ipc.msq
- monitored_instance:
- name: kern.ipc.msq
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect number of IPC message Queues"
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: kern.ipc.msq
- description: Enable or disable IPC message queue metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show statistics IPC messages statistics."
- labels: []
- metrics:
- - name: system.ipc_msq_queues
- description: Number of IPC Message Queues
- unit: "queues"
- chart_type: area
- dimensions:
- - name: queues
- - name: system.ipc_msq_messages
- description: Number of Messages in IPC Message Queues
- unit: "messages"
- chart_type: area
- dimensions:
- - name: messages
- - name: system.ipc_msq_size
- description: Size of IPC Message Queues
- unit: "bytes"
- chart_type: line
- dimensions:
- - name: allocated
- - name: used
- - meta:
- plugin_name: freebsd.plugin
- module_name: uptime
- monitored_instance:
- name: uptime
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Show period of time server is up."
- method_description: "The plugin calls `clock_gettime` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: vm.loadavg
- description: Enable or disable load average metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "How long the system is running."
- labels: []
- metrics:
- - name: system.uptime
- description: System Uptime
- unit: "seconds"
- chart_type: line
- dimensions:
- - name: uptime
- - meta:
- plugin_name: freebsd.plugin
- module_name: net.isr
- monitored_instance:
- name: net.isr
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "freebsd.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect information about system softnet stat."
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:net.isr]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: netisr
- description: Enable or disable general vision about softnet stat metrics.
- default_value: yes
- required: false
- - name: netisr per core
- description: Enable or disable softnet stat metric per core.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 1min_netdev_backlog_exceeded
- link: https://github.com/netdata/netdata/blob/master/health/health.d/softnet.conf
- metric: system.softnet_stat
- info: average number of dropped packets in the last minute due to exceeded net.core.netdev_max_backlog
- os: "linux"
- - name: 1min_netdev_budget_ran_outs
- link: https://github.com/netdata/netdata/blob/master/health/health.d/softnet.conf
- metric: system.softnet_stat
- info:
- average number of times ksoftirq ran out of sysctl net.core.netdev_budget or net.core.netdev_budget_usecs with work remaining over the last
- minute (this can be a cause for dropped packets)
- os: "linux"
- - name: 10min_netisr_backlog_exceeded
- link: https://github.com/netdata/netdata/blob/master/health/health.d/softnet.conf
- metric: system.softnet_stat
- info: average number of drops in the last minute due to exceeded sysctl net.route.netisr_maxqlen (this can be a cause for dropped packets)
- os: "freebsd"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show statistics about softnet stats."
- labels: []
- metrics:
- - name: system.softnet_stat
- description: System softnet_stat
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: dispatched
- - name: hybrid_dispatched
- - name: qdrops
- - name: queued
- - name: core
- description: ""
- labels: []
- metrics:
- - name: cpu.softnet_stat
- description: Per CPU netisr statistics
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: dispatched
- - name: hybrid_dispatched
- - name: qdrops
- - name: queued
- - meta:
- plugin_name: freebsd.plugin
- module_name: devstat
- monitored_instance:
- name: devstat
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "hard-drive.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect information per hard disk available on host."
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:kern.devstat]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: enable new disks detected at runtime
- description: Enable or disable possibility to detect new disks.
- default_value: auto
- required: false
- - name: performance metrics for pass devices
- description: Enable or disable metrics for disks with type `PASS`.
- default_value: auto
- required: false
- - name: total bandwidth for all disks
- description: Enable or disable total bandwidth metric for all disks.
- default_value: yes
- required: false
- - name: bandwidth for all disks
- description: Enable or disable bandwidth for all disks metric.
- default_value: auto
- required: false
- - name: operations for all disks
- description: Enable or disable operations for all disks metric.
- default_value: auto
- required: false
- - name: queued operations for all disks
- description: Enable or disable queued operations for all disks metric.
- default_value: auto
- required: false
- - name: utilization percentage for all disks
- description: Enable or disable utilization percentage for all disks metric.
- default_value: auto
- required: false
- - name: i/o time for all disks
- description: Enable or disable I/O time for all disks metric.
- default_value: auto
- required: false
- - name: average completed i/o time for all disks
- description: Enable or disable average completed I/O time for all disks metric.
- default_value: auto
- required: false
- - name: average completed i/o bandwidth for all disks
- description: Enable or disable average completed I/O bandwidth for all disks metric.
- default_value: auto
- required: false
- - name: average service time for all disks
- description: Enable or disable average service time for all disks metric.
- default_value: auto
- required: false
- - name: disable by default disks matching
- description: Do not create charts for disks listed.
- default_value: ""
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 10min_disk_utilization
- link: https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf
- metric: disk.util
- info: average percentage of time ${label:device} disk was busy over the last 10 minutes
- os: "linux freebsd"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics give a general vision about I/O events on disks."
- labels: []
- metrics:
- - name: system.io
- description: Disk I/O
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: io
- - name: out
- - name: disk
- description: ""
- labels: []
- metrics:
- - name: disk.io
- description: Disk I/O Bandwidth
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: reads
- - name: writes
- - name: frees
- - name: disk.ops
- description: Disk Completed I/O Operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: reads
- - name: writes
- - name: other
- - name: frees
- - name: disk.qops
- description: Disk Current I/O Operations
- unit: "operations"
- chart_type: line
- dimensions:
- - name: operations
- - name: disk.util
- description: Disk Utilization Time
- unit: "% of time working"
- chart_type: line
- dimensions:
- - name: utilization
- - name: disk.iotime
- description: Disk Total I/O Time
- unit: "milliseconds/s"
- chart_type: line
- dimensions:
- - name: reads
- - name: writes
- - name: other
- - name: frees
- - name: disk.await
- description: Average Completed I/O Operation Time
- unit: "milliseconds/operation"
- chart_type: line
- dimensions:
- - name: reads
- - name: writes
- - name: other
- - name: frees
- - name: disk.avgsz
- description: Average Completed I/O Operation Bandwidth
- unit: "KiB/operation"
- chart_type: area
- dimensions:
- - name: reads
- - name: writes
- - name: frees
- - name: disk.svctm
- description: Average Service Time
- unit: "milliseconds/operation"
- chart_type: line
- dimensions:
- - name: svctm
- - meta:
- plugin_name: freebsd.plugin
- module_name: net.inet.tcp.states
- monitored_instance:
- name: net.inet.tcp.states
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "network.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: ""
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: net.inet.tcp.states
- description: Enable or disable TCP state metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: tcp_connections
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_conn.conf
- metric: ipv4.tcpsock
- info: IPv4 TCP connections utilization
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "A counter for TCP connections."
- labels: []
- metrics:
- - name: ipv4.tcpsock
- description: IPv4 TCP Connections
- unit: "active connections"
- chart_type: line
- dimensions:
- - name: connections
- - meta:
- plugin_name: freebsd.plugin
- module_name: net.inet.tcp.stats
- monitored_instance:
- name: net.inet.tcp.stats
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "network.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect overall information about TCP connections."
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:net.inet.tcp.stats]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: ipv4 TCP packets
- description: Enable or disable ipv4 TCP packets metric.
- default_value: yes
- required: false
- - name: ipv4 TCP errors
- description: Enable or disable pv4 TCP errors metric.
- default_value: yes
- required: false
- - name: ipv4 TCP handshake issues
- description: Enable or disable ipv4 TCP handshake issue metric.
- default_value: yes
- required: false
- - name: TCP connection aborts
- description: Enable or disable TCP connection aborts metric.
- default_value: auto
- required: false
- - name: TCP out-of-order queue
- description: Enable or disable TCP out-of-order queue metric.
- default_value: auto
- required: false
- - name: TCP SYN cookies
- description: Enable or disable TCP SYN cookies metric.
- default_value: auto
- required: false
- - name: TCP listen issues
- description: Enable or disable TCP listen issues metric.
- default_value: auto
- required: false
- - name: ECN packets
- description: Enable or disable ECN packets metric.
- default_value: auto
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 1m_ipv4_tcp_resets_sent
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf
- metric: ipv4.tcphandshake
- info: average number of sent TCP RESETS over the last minute
- os: "linux"
- - name: 10s_ipv4_tcp_resets_sent
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf
- metric: ipv4.tcphandshake
- info:
- average number of sent TCP RESETS over the last 10 seconds. This can indicate a port scan, or that a service running on this host has
- crashed. Netdata will not send a clear notification for this alarm.
- os: "linux"
- - name: 1m_ipv4_tcp_resets_received
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf
- metric: ipv4.tcphandshake
- info: average number of received TCP RESETS over the last minute
- os: "linux freebsd"
- - name: 10s_ipv4_tcp_resets_received
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf
- metric: ipv4.tcphandshake
- info:
- average number of received TCP RESETS over the last 10 seconds. This can be an indication that a service this host needs has crashed.
- Netdata will not send a clear notification for this alarm.
- os: "linux freebsd"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show TCP connections statistics."
- labels: []
- metrics:
- - name: ipv4.tcppackets
- description: IPv4 TCP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.tcperrors
- description: IPv4 TCP Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InErrs
- - name: InCsumErrors
- - name: RetransSegs
- - name: ipv4.tcphandshake
- description: IPv4 TCP Handshake Issues
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: EstabResets
- - name: ActiveOpens
- - name: PassiveOpens
- - name: AttemptFails
- - name: ipv4.tcpconnaborts
- description: TCP Connection Aborts
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: baddata
- - name: userclosed
- - name: nomemory
- - name: timeout
- - name: linger
- - name: ipv4.tcpofo
- description: TCP Out-Of-Order Queue
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: inqueue
- - name: ipv4.tcpsyncookies
- description: TCP SYN Cookies
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: failed
- - name: ipv4.tcplistenissues
- description: TCP Listen Socket Issues
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: overflows
- - name: ipv4.ecnpkts
- description: IPv4 ECN Statistics
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InCEPkts
- - name: InECT0Pkts
- - name: InECT1Pkts
- - name: OutECT0Pkts
- - name: OutECT1Pkts
- - meta:
- plugin_name: freebsd.plugin
- module_name: net.inet.udp.stats
- monitored_instance:
- name: net.inet.udp.stats
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "network.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect information about UDP connections."
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:net.inet.udp.stats]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: ipv4 UDP packets
- description: Enable or disable ipv4 UDP packets metric.
- default_value: yes
- required: false
- - name: ipv4 UDP errors
- description: Enable or disable ipv4 UDP errors metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 1m_ipv4_udp_receive_buffer_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/udp_errors.conf
- metric: ipv4.udperrors
- info: average number of UDP receive buffer errors over the last minute
- os: "linux freebsd"
- - name: 1m_ipv4_udp_send_buffer_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/udp_errors.conf
- metric: ipv4.udperrors
- info: average number of UDP send buffer errors over the last minute
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show UDP connections statistics."
- labels: []
- metrics:
- - name: ipv4.udppackets
- description: IPv4 UDP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.udperrors
- description: IPv4 UDP Errors
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: InErrors
- - name: NoPorts
- - name: RcvbufErrors
- - name: InCsumErrors
- - name: IgnoredMulti
- - meta:
- plugin_name: freebsd.plugin
- module_name: net.inet.icmp.stats
- monitored_instance:
- name: net.inet.icmp.stats
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "network.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect information about ICMP traffic."
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:net.inet.icmp.stats]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: IPv4 ICMP packets
- description: Enable or disable IPv4 ICMP packets metric.
- default_value: yes
- required: false
- - name: IPv4 ICMP error
- description: Enable or disable IPv4 ICMP error metric.
- default_value: yes
- required: false
- - name: IPv4 ICMP messages
- description: Enable or disable IPv4 ICMP messages metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show ICMP connections statistics."
- labels: []
- metrics:
- - name: ipv4.icmp
- description: IPv4 ICMP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.icmp_errors
- description: IPv4 ICMP Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InErrors
- - name: OutErrors
- - name: InCsumErrors
- - name: ipv4.icmpmsg
- description: IPv4 ICMP Messages
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InEchoReps
- - name: OutEchoReps
- - name: InEchos
- - name: OutEchos
- - meta:
- plugin_name: freebsd.plugin
- module_name: net.inet.ip.stats
- monitored_instance:
- name: net.inet.ip.stats
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "network.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect IP stats"
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:net.inet.ip.stats]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: ipv4 packets
- description: Enable or disable IPv4 packets metric.
- default_value: yes
- required: false
- - name: ipv4 fragments sent
- description: Enable or disable IPv4 fragments sent metric.
- default_value: yes
- required: false
- - name: ipv4 fragments assembly
- description: Enable or disable IPv4 fragments assembly metric.
- default_value: yes
- required: false
- - name: ipv4 errors
- description: Enable or disable IPv4 errors metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show IPv4 connections statistics."
- labels: []
- metrics:
- - name: ipv4.packets
- description: IPv4 Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: forwarded
- - name: delivered
- - name: ipv4.fragsout
- description: IPv4 Fragments Sent
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: created
- - name: ipv4.fragsin
- description: IPv4 Fragments Reassembly
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: all
- - name: ipv4.errors
- description: IPv4 Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InDiscards
- - name: OutDiscards
- - name: InHdrErrors
- - name: OutNoRoutes
- - name: InAddrErrors
- - name: InUnknownProtos
- - meta:
- plugin_name: freebsd.plugin
- module_name: net.inet6.ip6.stats
- monitored_instance:
- name: net.inet6.ip6.stats
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "network.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect information abou IPv6 stats."
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:net.inet6.ip6.stats]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: ipv6 packets
- description: Enable or disable ipv6 packet metric.
- default_value: auto
- required: false
- - name: ipv6 fragments sent
- description: Enable or disable ipv6 fragments sent metric.
- default_value: auto
- required: false
- - name: ipv6 fragments assembly
- description: Enable or disable ipv6 fragments assembly metric.
- default_value: auto
- required: false
- - name: ipv6 errors
- description: Enable or disable ipv6 errors metric.
- default_value: auto
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show general information about IPv6 connections."
- labels: []
- metrics:
- - name: ipv6.packets
- description: IPv6 Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: forwarded
- - name: delivers
- - name: ipv6.fragsout
- description: IPv6 Fragments Sent
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: all
- - name: ipv6.fragsin
- description: IPv6 Fragments Reassembly
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: timeout
- - name: all
- - name: ipv6.errors
- description: IPv6 Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InDiscards
- - name: OutDiscards
- - name: InHdrErrors
- - name: InAddrErrors
- - name: InTruncatedPkts
- - name: InNoRoutes
- - name: OutNoRoutes
- - meta:
- plugin_name: freebsd.plugin
- module_name: net.inet6.icmp6.stats
- monitored_instance:
- name: net.inet6.icmp6.stats
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "network.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect information abou IPv6 ICMP"
- method_description: "The plugin calls `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:net.inet6.icmp6.stats]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: icmp
- description: Enable or disable ICMP metric.
- default_value: auto
- required: false
- - name: icmp redirects
- description: Enable or disable ICMP redirects metric.
- default_value: auto
- required: false
- - name: icmp errors
- description: Enable or disable ICMP errors metric.
- default_value: auto
- required: false
- - name: icmp echos
- description: Enable or disable ICMP echos metric.
- default_value: auto
- required: false
- - name: icmp router
- description: Enable or disable ICMP router metric.
- default_value: auto
- required: false
- - name: icmp neighbor
- description: Enable or disable ICMP neighbor metric.
- default_value: auto
- required: false
- - name: icmp types
- description: Enable or disable ICMP types metric.
- default_value: auto
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "Collect IPv6 ICMP traffic statistics."
- labels: []
- metrics:
- - name: ipv6.icmp
- description: IPv6 ICMP Messages
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.icmpredir
- description: IPv6 ICMP Redirects
- unit: "redirects/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.icmperrors
- description: IPv6 ICMP Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: InErrors
- - name: OutErrors
- - name: InCsumErrors
- - name: InDestUnreachs
- - name: InPktTooBigs
- - name: InTimeExcds
- - name: InParmProblems
- - name: OutDestUnreachs
- - name: OutTimeExcds
- - name: OutParmProblems
- - name: ipv6.icmpechos
- description: IPv6 ICMP Echo
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InEchos
- - name: OutEchos
- - name: InEchoReplies
- - name: OutEchoReplies
- - name: ipv6.icmprouter
- description: IPv6 Router Messages
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InSolicits
- - name: OutSolicits
- - name: InAdvertisements
- - name: OutAdvertisements
- - name: ipv6.icmpneighbor
- description: IPv6 Neighbor Messages
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InSolicits
- - name: OutSolicits
- - name: InAdvertisements
- - name: OutAdvertisements
- - name: ipv6.icmptypes
- description: IPv6 ICMP Types
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InType1
- - name: InType128
- - name: InType129
- - name: InType136
- - name: OutType1
- - name: OutType128
- - name: OutType129
- - name: OutType133
- - name: OutType135
- - name: OutType143
- - meta:
- plugin_name: freebsd.plugin
- module_name: ipfw
- monitored_instance:
- name: ipfw
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "firewall.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect information about FreeBSD firewall."
- method_description: "The plugin uses RAW socket to communicate with kernel and collect data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:ipfw]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: counters for static rules
- description: Enable or disable counters for static rules metric.
- default_value: yes
- required: false
- - name: number of dynamic rules
- description: Enable or disable number of dynamic rules metric.
- default_value: yes
- required: false
- - name: allocated memory
- description: Enable or disable allocated memory metric.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "Theese metrics show FreeBSD firewall statistics."
- labels: []
- metrics:
- - name: ipfw.mem
- description: Memory allocated by rules
- unit: "bytes"
- chart_type: stacked
- dimensions:
- - name: dynamic
- - name: static
- - name: ipfw.packets
- description: Packets
- unit: "packets/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per static rule
- - name: ipfw.bytes
- description: Bytes
- unit: "bytes/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per static rule
- - name: ipfw.active
- description: Active rules
- unit: "rules"
- chart_type: stacked
- dimensions:
- - name: a dimension per dynamic rule
- - name: ipfw.expired
- description: Expired rules
- unit: "rules"
- chart_type: stacked
- dimensions:
- - name: a dimension per dynamic rule
- - meta:
- plugin_name: freebsd.plugin
- module_name: getifaddrs
- monitored_instance:
- name: getifaddrs
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "network.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect traffic per network interface."
- method_description: "The plugin calls `getifaddrs` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:getifaddrs]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: enable new interfaces detected at runtime
- description: Enable or disable possibility to discover new interface after plugin starts.
- default_value: auto
- required: false
- - name: total bandwidth for physical interfaces
- description: Enable or disable total bandwidth for physical interfaces metric.
- default_value: auto
- required: false
- - name: total packets for physical interfaces
- description: Enable or disable total packets for physical interfaces metric.
- default_value: auto
- required: false
- - name: total bandwidth for ipv4 interface
- description: Enable or disable total bandwidth for IPv4 interface metric.
- default_value: auto
- required: false
- - name: total bandwidth for ipv6 interfaces
- description: Enable or disable total bandwidth for ipv6 interfaces metric.
- default_value: auto
- required: false
- - name: bandwidth for all interfaces
- description: Enable or disable bandwidth for all interfaces metric.
- default_value: auto
- required: false
- - name: packets for all interfaces
- description: Enable or disable packets for all interfaces metric.
- default_value: auto
- required: false
- - name: errors for all interfaces
- description: Enable or disable errors for all interfaces metric.
- default_value: auto
- required: false
- - name: drops for all interfaces
- description: Enable or disable drops for all interfaces metric.
- default_value: auto
- required: false
- - name: collisions for all interface
- description: Enable or disable collisions for all interface metric.
- default_value: auto
- required: false
- - name: disable by default interfaces matching
- description: Do not display data for intterfaces listed.
- default_value: lo*
- required: false
- - name: set physical interfaces for system.net
- description: Do not show network traffic for listed interfaces.
- default_value: igb* ix* cxl* em* ixl* ixlv* bge* ixgbe* vtnet* vmx* re* igc* dwc*
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: interface_speed
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.net
- info: network interface ${label:device} current speed
- os: "*"
- - name: inbound_packets_dropped_ratio
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.drops
- info: ratio of inbound dropped packets for the network interface ${label:device} over the last 10 minutes
- os: "*"
- - name: outbound_packets_dropped_ratio
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.drops
- info: ratio of outbound dropped packets for the network interface ${label:device} over the last 10 minutes
- os: "*"
- - name: 1m_received_packets_rate
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.packets
- info: average number of packets received by the network interface ${label:device} over the last minute
- os: "linux freebsd"
- - name: 10s_received_packets_storm
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.packets
- info: ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute
- os: "linux freebsd"
- - name: interface_inbound_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.errors
- info: number of inbound errors for the network interface ${label:device} in the last 10 minutes
- os: "freebsd"
- - name: interface_outbound_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.errors
- info: number of outbound errors for the network interface ${label:device} in the last 10 minutes
- os: "freebsd"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "General overview about network traffic."
- labels: []
- metrics:
- - name: system.net
- description: Network Traffic
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: system.packets
- description: Network Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: multicast_received
- - name: multicast_sent
- - name: system.ipv4
- description: IPv4 Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: system.ipv6
- description: IPv6 Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: network device
- description: ""
- labels: []
- metrics:
- - name: net.net
- description: Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: net.packets
- description: Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: multicast_received
- - name: multicast_sent
- - name: net.errors
- description: Interface Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: inbound
- - name: outbound
- - name: net.drops
- description: Interface Drops
- unit: "drops/s"
- chart_type: line
- dimensions:
- - name: inbound
- - name: outbound
- - name: net.events
- description: Network Interface Events
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: collisions
- - meta:
- plugin_name: freebsd.plugin
- module_name: getmntinfo
- monitored_instance:
- name: getmntinfo
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "hard-drive.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect information per mount point."
- method_description: "The plugin calls `getmntinfo` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:getmntinfo]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: enable new mount points detected at runtime
- description: Cheeck new mount points during runtime.
- default_value: auto
- required: false
- - name: space usage for all disks
- description: Enable or disable space usage for all disks metric.
- default_value: auto
- required: false
- - name: inodes usage for all disks
- description: Enable or disable inodes usage for all disks metric.
- default_value: auto
- required: false
- - name: exclude space metrics on paths
- description: Do not show metrics for listed paths.
- default_value: /proc/*
- required: false
- - name: exclude space metrics on filesystems
- description: Do not monitor listed filesystems.
- default_value: autofs procfs subfs devfs none
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: disk_space_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf
- metric: disk.space
- info: disk ${label:mount_point} space utilization
- os: "linux freebsd"
- - name: disk_inode_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf
- metric: disk.inodes
- info: disk ${label:mount_point} inode utilization
- os: "linux freebsd"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: mount point
- description: "These metrics show detailss about mount point usages."
- labels: []
- metrics:
- - name: disk.space
- description: Disk Space Usage for {mounted dir} [{mounted filesystem}]
- unit: "GiB"
- chart_type: stacked
- dimensions:
- - name: avail
- - name: used
- - name: reserved_for_root
- - name: disk.inodes
- description: Disk Files (inodes) Usage for {mounted dir} [{mounted filesystem}]
- unit: "inodes"
- chart_type: stacked
- dimensions:
- - name: avail
- - name: used
- - name: reserved_for_root
- - meta:
- plugin_name: freebsd.plugin
- module_name: zfs
- monitored_instance:
- name: zfs
- link: "https://www.freebsd.org/"
- categories:
- - data-collection.freebsd
- icon_filename: "filesystem.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Collect metrics for ZFS filesystem"
- method_description: "The plugin uses `sysctl` function to collect necessary data."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freebsd:zfs_arcstats]"
- description: "The netdata main configuration file."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: show zero charts
- description: Do not show charts with zero metrics.
- default_value: no
- required: false
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: zfs_memory_throttle
- link: https://github.com/netdata/netdata/blob/master/health/health.d/zfs.conf
- metric: zfs.memory_ops
- info: number of times ZFS had to limit the ARC growth in the last 10 minutes
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics show detailed information about ZFS filesystem."
- labels: []
- metrics:
- - name: zfs.arc_size
- description: ZFS ARC Size
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: arcsz
- - name: target
- - name: min
- - name: max
- - name: zfs.l2_size
- description: ZFS L2 ARC Size
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: actual
- - name: size
- - name: zfs.reads
- description: ZFS Reads
- unit: "reads/s"
- chart_type: area
- dimensions:
- - name: arc
- - name: demand
- - name: prefetch
- - name: metadata
- - name: l2
- - name: zfs.bytes
- description: ZFS ARC L2 Read/Write Rate
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: zfs.hits
- description: ZFS ARC Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.hits_rate
- description: ZFS ARC Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.dhits
- description: ZFS Demand Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.dhits_rate
- description: ZFS Demand Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.phits
- description: ZFS Prefetch Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.phits_rate
- description: ZFS Prefetch Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.mhits
- description: ZFS Metadata Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.mhits_rate
- description: ZFS Metadata Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.l2hits
- description: ZFS L2 Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.l2hits_rate
- description: ZFS L2 Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.list_hits
- description: ZFS List Hits
- unit: "hits/s"
- chart_type: area
- dimensions:
- - name: mfu
- - name: mfu_ghost
- - name: mru
- - name: mru_ghost
- - name: zfs.arc_size_breakdown
- description: ZFS ARC Size Breakdown
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: recent
- - name: frequent
- - name: zfs.memory_ops
- description: ZFS Memory Operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: throttled
- - name: zfs.important_ops
- description: ZFS Important Operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: evict_skip
- - name: deleted
- - name: mutex_miss
- - name: hash_collisions
- - name: zfs.actual_hits
- description: ZFS Actual Cache Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.actual_hits_rate
- description: ZFS Actual Cache Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.demand_data_hits
- description: ZFS Data Demand Efficiency
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.demand_data_hits_rate
- description: ZFS Data Demand Efficiency Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.prefetch_data_hits
- description: ZFS Data Prefetch Efficiency
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.prefetch_data_hits_rate
- description: ZFS Data Prefetch Efficiency Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.hash_elements
- description: ZFS ARC Hash Elements
- unit: "elements"
- chart_type: line
- dimensions:
- - name: current
- - name: max
- - name: zfs.hash_chains
- description: ZFS ARC Hash Chains
- unit: "chains"
- chart_type: line
- dimensions:
- - name: current
- - name: max
- - name: zfs.trim_bytes
- description: Successfully TRIMmed bytes
- unit: "bytes"
- chart_type: line
- dimensions:
- - name: TRIMmed
- - name: zfs.trim_requests
- description: TRIM requests
- unit: "requests"
- chart_type: line
- dimensions:
- - name: successful
- - name: failed
- - name: unsupported
diff --git a/collectors/freeipmi.plugin/Makefile.am b/collectors/freeipmi.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/freeipmi.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/freeipmi.plugin/metadata.yaml b/collectors/freeipmi.plugin/metadata.yaml
deleted file mode 100644
index f8c75c2cb..000000000
--- a/collectors/freeipmi.plugin/metadata.yaml
+++ /dev/null
@@ -1,347 +0,0 @@
-plugin_name: freeipmi.plugin
-modules:
- - meta:
- plugin_name: freeipmi.plugin
- module_name: freeipmi
- monitored_instance:
- name: Intelligent Platform Management Interface (IPMI)
- link: "https://en.wikipedia.org/wiki/Intelligent_Platform_Management_Interface"
- categories:
- - data-collection.hardware-devices-and-sensors
- icon_filename: "netdata.png"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - sensors
- - ipmi
- - freeipmi
- - ipmimonitoring
- most_popular: true
- overview:
- data_collection:
- metrics_description: |
- "Monitor enterprise server sensor readings, event log entries, and hardware statuses to ensure reliable server operations."
- method_description: |
- The plugin uses open source library IPMImonitoring to communicate with sensors.
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "The plugin needs setuid."
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: "Linux kernel module for IPMI can create big overhead."
- setup:
- prerequisites:
- list:
- - title: Install freeipmi.plugin
- description: |
- When using our official DEB/RPM packages, the FreeIPMI plugin is included in a separate package named `netdata-plugin-freeipmi` which needs to be manually installed using your system package manager. It is not installed automatically due to the large number of dependencies it requires.
-
- When using a static build of Netdata, the FreeIPMI plugin will be included and installed automatically, though you will still need to have FreeIPMI installed on your system to be able to use the plugin.
-
- When using a local build of Netdata, you need to ensure that the FreeIPMI development packages (typically called `libipmimonitoring-dev`, `libipmimonitoring-devel`, or `freeipmi-devel`) are installed when building Netdata.
- - title: Preliminary actions
- description: |
- If you have not previously used IPMI on your system, you will probably need to run the `ipmimonitoring` command as root
- to initialize IPMI settings so that the Netdata plugin works correctly. It should return information about available sensors on the system.
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:freeipmi]"
- options:
- description: |
- The configuration is set using command line options:
-
- ```
- # netdata.conf
- [plugin:freeipmi]
- command options = opt1 opt2 ... optN
- ```
-
- To display a help message listing the available command line options:
-
- ```bash
- ./usr/libexec/netdata/plugins.d/freeipmi.plugin --help
- ```
- folding:
- title: "Command options"
- enabled: true
- list:
- - name: SECONDS
- description: Data collection frequency.
- default_value: ""
- required: false
- - name: debug
- description: Enable verbose output.
- default_value: disabled
- required: false
- - name: no-sel
- description: Disable System Event Log (SEL) collection.
- default_value: disabled
- required: false
- - name: reread-sdr-cache
- description: Re-read SDR cache on every iteration.
- default_value: disabled
- required: false
- - name: interpret-oem-data
- description: Attempt to parse OEM data.
- default_value: disabled
- required: false
- - name: assume-system-event-record
- description: treat illegal SEL events records as normal.
- default_value: disabled
- required: false
- - name: ignore-non-interpretable-sensors
- description: Do not read sensors that cannot be interpreted.
- default_value: disabled
- required: false
- - name: bridge-sensors
- description: Bridge sensors not owned by the BMC.
- default_value: disabled
- required: false
- - name: shared-sensors
- description: Enable shared sensors if found.
- default_value: disabled
- required: false
- - name: no-discrete-reading
- description: Do not read sensors if their event/reading type code is invalid.
- default_value: enabled
- required: false
- - name: ignore-scanning-disabled
- description: Ignore the scanning bit and read sensors no matter what.
- default_value: disabled
- required: false
- - name: assume-bmc-owner
- description: Assume the BMC is the sensor owner no matter what (usually bridging is required too).
- default_value: disabled
- required: false
- - name: hostname HOST
- description: Remote IPMI hostname or IP address.
- default_value: local
- required: false
- - name: username USER
- description: Username that will be used when connecting to the remote host.
- default_value: ""
- required: false
- - name: password PASS
- description: Password that will be used when connecting to the remote host.
- default_value: ""
- required: false
- - name: noauthcodecheck / no-auth-code-check
- description: Don't check the authentication codes returned.
- default_value: ""
- required: false
- - name: driver-type IPMIDRIVER
- description: Specify the driver type to use instead of doing an auto selection. The currently available outofband drivers are LAN and LAN_2_0, which perform IPMI 1.5 and IPMI 2.0 respectively. The currently available inband drivers are KCS, SSIF, OPENIPMI and SUNBMC.
- default_value: ""
- required: false
- - name: sdr-cache-dir PATH
- description: SDR cache files directory.
- default_value: /tmp
- required: false
- - name: sensor-config-file FILE
- description: Sensors configuration filename.
- default_value: system default
- required: false
- - name: sel-config-file FILE
- description: SEL configuration filename.
- default_value: system default
- required: false
- - name: ignore N1,N2,N3,...
- description: Sensor IDs to ignore.
- default_value: ""
- required: false
- - name: ignore-status N1,N2,N3,...
- description: Sensor IDs to ignore status (nominal/warning/critical).
- default_value: ""
- required: false
- - name: -v
- description: Print version and exit.
- default_value: ""
- required: false
- - name: --help
- description: Print usage message and exit.
- default_value: ""
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Decrease data collection frequency
- description: Basic example decreasing data collection frequency. The minimum `update every` is 5 (enforced internally by the plugin). IPMI is slow and CPU hungry. So, once every 5 seconds is pretty acceptable.
- config: |
- [plugin:freeipmi]
- update every = 10
- folding:
- enabled: false
- - name: Disable SEL collection
- description: Append to `command options =` the options you need.
- config: |
- [plugin:freeipmi]
- command options = no-sel
- - name: Ignore specific sensors
- description: |
- Specific sensor IDs can be excluded from freeipmi tools by editing `/etc/freeipmi/freeipmi.conf` and setting the IDs to be ignored at `ipmi-sensors-exclude-record-ids`.
-
- **However this file is not used by `libipmimonitoring`** (the library used by Netdata's `freeipmi.plugin`).
-
- To find the IDs to ignore, run the command `ipmimonitoring`. The first column is the wanted ID:
-
- ID | Name | Type | State | Reading | Units | Event
- 1 | Ambient Temp | Temperature | Nominal | 26.00 | C | 'OK'
- 2 | Altitude | Other Units Based Sensor | Nominal | 480.00 | ft | 'OK'
- 3 | Avg Power | Current | Nominal | 100.00 | W | 'OK'
- 4 | Planar 3.3V | Voltage | Nominal | 3.29 | V | 'OK'
- 5 | Planar 5V | Voltage | Nominal | 4.90 | V | 'OK'
- 6 | Planar 12V | Voltage | Nominal | 11.99 | V | 'OK'
- 7 | Planar VBAT | Voltage | Nominal | 2.95 | V | 'OK'
- 8 | Fan 1A Tach | Fan | Nominal | 3132.00 | RPM | 'OK'
- 9 | Fan 1B Tach | Fan | Nominal | 2150.00 | RPM | 'OK'
- 10 | Fan 2A Tach | Fan | Nominal | 2494.00 | RPM | 'OK'
- 11 | Fan 2B Tach | Fan | Nominal | 1825.00 | RPM | 'OK'
- 12 | Fan 3A Tach | Fan | Nominal | 3538.00 | RPM | 'OK'
- 13 | Fan 3B Tach | Fan | Nominal | 2625.00 | RPM | 'OK'
- 14 | Fan 1 | Entity Presence | Nominal | N/A | N/A | 'Entity Present'
- 15 | Fan 2 | Entity Presence | Nominal | N/A | N/A | 'Entity Present'
- ...
-
- `freeipmi.plugin` supports the option `ignore` that accepts a comma separated list of sensor IDs to ignore. To configure it set on `netdata.conf`:
- config: |
- [plugin:freeipmi]
- command options = ignore 1,2,3,4,...
- troubleshooting:
- problems:
- list:
- - name: Debug Mode
- description: |
- You can run `freeipmi.plugin` with the debug option enabled, to troubleshoot issues with it. The output should give you clues as to why the collector isn't working.
-
- - Navigate to the `plugins.d` directory, usually at `/usr/libexec/netdata/plugins.d/`. If that's not the case on your system, open `netdata.conf` and look for the `plugins` setting under `[directories]`.
-
- ```bash
- cd /usr/libexec/netdata/plugins.d/
- ```
-
- - Switch to the `netdata` user.
-
- ```bash
- sudo -u netdata -s
- ```
-
- - Run the `freeipmi.plugin` in debug mode:
-
- ```bash
- ./freeipmi.plugin 5 debug
- ```
- - name: kimpi0 CPU usage
- description: |
- There have been reports that kipmi is showing increased CPU when the IPMI is queried. To lower the CPU consumption of the system you can issue this command:
-
- ```sh
- echo 10 > /sys/module/ipmi_si/parameters/kipmid_max_busy_us
- ```
-
- You can also permanently set the above setting by creating the file `/etc/modprobe.d/ipmi.conf` with this content:
-
- ```sh
- # prevent kipmi from consuming 100% CPU
- options ipmi_si kipmid_max_busy_us=10
- ```
-
- This instructs the kernel IPMI module to pause for a tick between checking IPMI. Querying IPMI will be a lot slower now (e.g. several seconds for IPMI to respond), but `kipmi` will not use any noticeable CPU.
-
- You can also use a higher number (this is the number of microseconds to poll IPMI for a response, before waiting for a tick).
- alerts:
- - name: ipmi_sensor_state
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ipmi.conf
- metric: ipmi.sensor_state
- info: IPMI sensor ${label:sensor} (${label:component}) state
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: |
- The plugin does a speed test when it starts, to find out the duration needed by the IPMI processor to respond. Depending on the speed of your IPMI processor, charts may need several seconds to show up on the dashboard.
- availability: []
- scopes:
- - name: global
- description: These metrics refer to the entire monitored application.
- labels: []
- metrics:
- - name: ipmi.sel
- description: IPMI Events
- unit: "events"
- chart_type: area
- dimensions:
- - name: events
- - name: sensor
- description: ""
- labels:
- - name: sensor
- description: The sensor name
- - name: type
- description: One of 45 recognized sensor types (Battery, Voltage...)
- - name: component
- description: One of 25 recognized components (Processor, Peripheral).
- metrics:
- - name: ipmi.sensor_state
- description: IPMI Sensors State
- unit: "state"
- chart_type: line
- dimensions:
- - name: nominal
- - name: critical
- - name: warning
- - name: unknown
- - name: ipmi.sensor_temperature_c
- description: IPMI Sensor Temperature Celsius
- unit: "Celsius"
- chart_type: line
- dimensions:
- - name: temperature
- - name: ipmi.sensor_temperature_f
- description: IPMI Sensor Temperature Fahrenheit
- unit: "Fahrenheit"
- chart_type: line
- dimensions:
- - name: temperature
- - name: ipmi.sensor_voltage
- description: IPMI Sensor Voltage
- unit: "Volts"
- chart_type: line
- dimensions:
- - name: voltage
- - name: ipmi.sensor_ampere
- description: IPMI Sensor Current
- unit: "Amps"
- chart_type: line
- dimensions:
- - name: ampere
- - name: ipmi.sensor_fan_speed
- description: IPMI Sensor Fans Speed
- unit: "RPM"
- chart_type: line
- dimensions:
- - name: rotations
- - name: ipmi.sensor_power
- description: IPMI Sensor Power
- unit: "Watts"
- chart_type: line
- dimensions:
- - name: power
- - name: ipmi.sensor_reading_percent
- description: IPMI Sensor Reading Percentage
- unit: "%"
- chart_type: line
- dimensions:
- - name: percentage
diff --git a/collectors/idlejitter.plugin/Makefile.am b/collectors/idlejitter.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/idlejitter.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/ioping.plugin/Makefile.am b/collectors/ioping.plugin/Makefile.am
deleted file mode 100644
index a9cd7c4f1..000000000
--- a/collectors/ioping.plugin/Makefile.am
+++ /dev/null
@@ -1,24 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-CLEANFILES = \
- ioping.plugin \
- $(NULL)
-
-include $(top_srcdir)/build/subst.inc
-SUFFIXES = .in
-
-dist_plugins_SCRIPTS = \
- ioping.plugin \
- $(NULL)
-
-dist_noinst_DATA = \
- ioping.plugin.in \
- README.md \
- $(NULL)
-
-dist_libconfig_DATA = \
- ioping.conf \
- $(NULL)
diff --git a/collectors/ioping.plugin/metadata.yaml b/collectors/ioping.plugin/metadata.yaml
deleted file mode 100644
index e3ec96162..000000000
--- a/collectors/ioping.plugin/metadata.yaml
+++ /dev/null
@@ -1,101 +0,0 @@
-plugin_name: ioping.plugin
-modules:
- - meta:
- plugin_name: ioping.plugin
- module_name: ioping.plugin
- monitored_instance:
- name: IOPing
- link: "https://github.com/koct9i/ioping"
- categories:
- - data-collection.synthetic-checks
- icon_filename: "syslog.png"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor IOPing metrics for efficient disk I/O latency tracking. Keep track of read/write speeds, latency, and error rates for optimized disk operations."
- method_description: "Plugin uses `ioping` command."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: Install ioping
- description: |
- You can install the command by passing the argument `install` to the plugin (`/usr/libexec/netdata/plugins.d/ioping.plugin install`).
- configuration:
- file:
- name: "ioping.conf"
- description: "File with options to specify hardware to monitor and arguments for ioping command."
- options:
- description: ""
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Data collection frequency.
- default_value: 1s
- required: false
- - name: destination
- description: The directory/file/device to ioping.
- default_value: ""
- required: true
- - name: request_size
- description: The request size in bytes to ioping the destination (symbolic modifiers are supported)
- default_value: 4k
- required: false
- - name: ioping_opts
- description: Options passed to `ioping` commands.
- default_value: -T 1000000
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Basic Configuration
- description: This example has the minimum configuration necessary to have the plugin running.
- config: |
- destination="/dev/sda"
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: ioping_disk_latency
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ioping.conf
- metric: ioping.latency
- info: average I/O latency over the last 10 seconds
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: disk
- description: ""
- labels: []
- metrics:
- - name: ioping.latency
- description: Read Latency
- unit: "microseconds"
- chart_type: line
- dimensions:
- - name: latency
diff --git a/collectors/log2journal/Makefile.am b/collectors/log2journal/Makefile.am
deleted file mode 100644
index b13d2160b..000000000
--- a/collectors/log2journal/Makefile.am
+++ /dev/null
@@ -1,17 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- tests.sh \
- README.md \
- tests.d/* \
- $(NULL)
-
-log2journalconfigdir=$(libconfigdir)/log2journal.d
-dist_log2journalconfig_DATA = \
- log2journal.d/nginx-combined.yaml \
- log2journal.d/nginx-json.yaml \
- log2journal.d/default.yaml \
- $(NULL)
diff --git a/collectors/log2journal/README.md b/collectors/log2journal/README.md
deleted file mode 100644
index 16ccc033c..000000000
--- a/collectors/log2journal/README.md
+++ /dev/null
@@ -1,912 +0,0 @@
-
-# log2journal
-
-`log2journal` and `systemd-cat-native` can be used to convert a structured log file, such as the ones generated by web servers, into `systemd-journal` entries.
-
-By combining these tools you can create advanced log processing pipelines sending any kind of structured text logs to systemd-journald. This is a simple, but powerful and efficient way to handle log processing.
-
-The process involves the usual piping of shell commands, to get and process the log files in realtime.
-
-The result is like this: nginx logs into systemd-journal:
-
-![image](https://github.com/netdata/netdata/assets/2662304/16b471ff-c5a1-4fcc-bcd5-83551e089f6c)
-
-
-The overall process looks like this:
-
-```bash
-tail -F /var/log/nginx/*.log |\ # outputs log lines
- log2journal 'PATTERN' |\ # outputs Journal Export Format
- systemd-cat-native # send to local/remote journald
-```
-
-These are the steps:
-
-1. `tail -F /var/log/nginx/*.log`<br/>this command will tail all `*.log` files in `/var/log/nginx/`. We use `-F` instead of `-f` to ensure that files will still be tailed after log rotation.
-2. `log2joural` is a Netdata program. It reads log entries and extracts fields, according to the PCRE2 pattern it accepts. It can also apply some basic operations on the fields, like injecting new fields or duplicating existing ones or rewriting their values. The output of `log2journal` is in Systemd Journal Export Format, and it looks like this:
- ```bash
- KEY1=VALUE1 # << start of the first log line
- KEY2=VALUE2
- # << log lines separator
- KEY1=VALUE1 # << start of the second log line
- KEY2=VALUE2
- ```
-3. `systemd-cat-native` is a Netdata program. I can send the logs to a local `systemd-journald` (journal namespaces supported), or to a remote `systemd-journal-remote`.
-
-
-## Processing pipeline
-
-The sequence of processing in Netdata's `log2journal` is designed to methodically transform and prepare log data for export in the systemd Journal Export Format. This transformation occurs through a pipeline of stages, each with a specific role in processing the log entries. Here's a description of each stage in the sequence:
-
-1. **Input**<br/>
- The tool reads one log line at a time from the input source. It supports different input formats such as JSON, logfmt, and free-form logs defined by PCRE2 patterns.
-
-2. **Extract Fields and Values**<br/>
- Based on the input format (JSON, logfmt, or custom pattern), it extracts fields and their values from each log line. In the case of JSON and logfmt, it automatically extracts all fields. For custom patterns, it uses PCRE2 regular expressions, and fields are extracted based on sub-expressions defined in the pattern.
-
-3. **Transliteration**<br/>
- Extracted fields are transliterated to the limited character set accepted by systemd-journal: capitals A-Z, digits 0-9, underscores.
-
-4. **Apply Optional Prefix**<br/>
- If a prefix is specified, it is added to all keys. This happens before any other processing so that all subsequent matches and manipulations take the prefix into account.
-
-5. **Rename Fields**<br/>
- Renames fields as specified in the configuration. This is used to change the names of the fields to match desired or required naming conventions.
-
-6. **Inject New Fields**<br/>
- New fields are injected into the log data. This can include constants or values derived from other fields, using variable substitution.
-
-7. **Rewrite Field Values**<br/>
- Applies rewriting rules to alter the values of the fields. This can involve complex transformations, including regular expressions and variable substitutions. The rewrite rules can also inject new fields into the data.
-
-8. **Filter Fields**<br/>
- Fields are filtered based on include and exclude patterns. This stage selects which fields are to be sent to the journal, allowing for selective logging.
-
-9. **Output**<br/>
- Finally, the processed log data is output in the Journal Export Format. This format is compatible with systemd's journaling system and can be sent to local or remote systemd journal systems, by piping the output of `log2journal` to `systemd-cat-native`.
-
-This pipeline ensures a flexible and comprehensive approach to log processing, allowing for a wide range of modifications and customizations to fit various logging requirements. Each stage builds upon the previous one, enabling complex log transformations and enrichments before the data is exported to the systemd journal.
-
-## Real-life example
-
-We have an nginx server logging in this standard combined log format:
-
-```bash
- log_format combined '$remote_addr - $remote_user [$time_local] '
- '"$request" $status $body_bytes_sent '
- '"$http_referer" "$http_user_agent"';
-```
-
-### Extracting fields with a pattern
-
-First, let's find the right pattern for `log2journal`. We ask ChatGPT:
-
-```
-My nginx log uses this log format:
-
-log_format access '$remote_addr - $remote_user [$time_local] '
- '"$request" $status $body_bytes_sent '
- '"$http_referer" "$http_user_agent"';
-
-I want to use `log2joural` to convert this log for systemd-journal.
-`log2journal` accepts a PCRE2 regular expression, using the named groups
-in the pattern as the journal fields to extract from the logs.
-
-Please give me the PCRE2 pattern to extract all the fields from my nginx
-log files.
-```
-
-ChatGPT replies with this:
-
-```regexp
- (?x) # Enable PCRE2 extended mode
- ^
- (?<remote_addr>[^ ]+) \s - \s
- (?<remote_user>[^ ]+) \s
- \[
- (?<time_local>[^\]]+)
- \]
- \s+ "
- (?<request>
- (?<request_method>[A-Z]+) \s+
- (?<request_uri>[^ ]+) \s+
- (?<server_protocol>[^"]+)
- )
- " \s+
- (?<status>\d+) \s+
- (?<body_bytes_sent>\d+) \s+
- "(?<http_referer>[^"]*)" \s+
- "(?<http_user_agent>[^"]*)"
-```
-
-Let's see what the above says:
-
-1. `(?x)`: enable PCRE2 extended mode. In this mode spaces and newlines in the pattern are ignored. To match a space you have to use `\s`. This mode allows us to split the pattern is multiple lines and add comments to it.
-1. `^`: match the beginning of the line
-2. `(?<remote_addr[^ ]+)`: match anything up to the first space (`[^ ]+`), and name it `remote_addr`.
-3. `\s`: match a space
-4. `-`: match a hyphen
-5. and so on...
-
-We edit `nginx.yaml` and add it, like this:
-
-```yaml
-pattern: |
- (?x) # Enable PCRE2 extended mode
- ^
- (?<remote_addr>[^ ]+) \s - \s
- (?<remote_user>[^ ]+) \s
- \[
- (?<time_local>[^\]]+)
- \]
- \s+ "
- (?<request>
- (?<request_method>[A-Z]+) \s+
- (?<request_uri>[^ ]+) \s+
- (?<server_protocol>[^"]+)
- )
- " \s+
- (?<status>\d+) \s+
- (?<body_bytes_sent>\d+) \s+
- "(?<http_referer>[^"]*)" \s+
- "(?<http_user_agent>[^"]*)"
-```
-
-Let's test it with a sample line (instead of `tail`):
-
-```bash
-# echo '1.2.3.4 - - [19/Nov/2023:00:24:43 +0000] "GET /index.html HTTP/1.1" 200 4172 104 0.001 "-" "Go-http-client/1.1"' | log2journal -f nginx.yaml
-BODY_BYTES_SENT=4172
-HTTP_REFERER=-
-HTTP_USER_AGENT=Go-http-client/1.1
-REMOTE_ADDR=1.2.3.4
-REMOTE_USER=-
-REQUEST=GET /index.html HTTP/1.1
-REQUEST_METHOD=GET
-REQUEST_URI=/index.html
-SERVER_PROTOCOL=HTTP/1.1
-STATUS=200
-TIME_LOCAL=19/Nov/2023:00:24:43 +0000
-
-```
-
-As you can see, it extracted all the fields and made them capitals, as systemd-journal expects them.
-
-### Prefixing field names
-
-To make sure the fields are unique for nginx and do not interfere with other applications, we should prefix them with `NGINX_`:
-
-```yaml
-pattern: |
- (?x) # Enable PCRE2 extended mode
- ^
- (?<remote_addr>[^ ]+) \s - \s
- (?<remote_user>[^ ]+) \s
- \[
- (?<time_local>[^\]]+)
- \]
- \s+ "
- (?<request>
- (?<request_method>[A-Z]+) \s+
- (?<request_uri>[^ ]+) \s+
- (?<server_protocol>[^"]+)
- )
- " \s+
- (?<status>\d+) \s+
- (?<body_bytes_sent>\d+) \s+
- "(?<http_referer>[^"]*)" \s+
- "(?<http_user_agent>[^"]*)"
-
-prefix: 'NGINX_' # <<< we added this
-```
-
-And let's try it:
-
-```bash
-# echo '1.2.3.4 - - [19/Nov/2023:00:24:43 +0000] "GET /index.html HTTP/1.1" 200 4172 "-" "Go-http-client/1.1"' | log2journal -f nginx.yaml
-NGINX_BODY_BYTES_SENT=4172
-NGINX_HTTP_REFERER=-
-NGINX_HTTP_USER_AGENT=Go-http-client/1.1
-NGINX_REMOTE_ADDR=1.2.3.4
-NGINX_REMOTE_USER=-
-NGINX_REQUEST=GET /index.html HTTP/1.1
-NGINX_REQUEST_METHOD=GET
-NGINX_REQUEST_URI=/index.html
-NGINX_SERVER_PROTOCOL=HTTP/1.1
-NGINX_STATUS=200
-NGINX_TIME_LOCAL=19/Nov/2023:00:24:43 +0000
-
-```
-
-### Renaming fields
-
-Now, all fields start with `NGINX_` but we want `NGINX_REQUEST` to be the `MESSAGE` of the log line, as we will see it by default in `journalctl` and the Netdata dashboard. Let's rename it:
-
-```yaml
-pattern: |
- (?x) # Enable PCRE2 extended mode
- ^
- (?<remote_addr>[^ ]+) \s - \s
- (?<remote_user>[^ ]+) \s
- \[
- (?<time_local>[^\]]+)
- \]
- \s+ "
- (?<request>
- (?<request_method>[A-Z]+) \s+
- (?<request_uri>[^ ]+) \s+
- (?<server_protocol>[^"]+)
- )
- " \s+
- (?<status>\d+) \s+
- (?<body_bytes_sent>\d+) \s+
- "(?<http_referer>[^"]*)" \s+
- "(?<http_user_agent>[^"]*)"
-
-prefix: 'NGINX_'
-
-rename: # <<< we added this
- - new_key: MESSAGE # <<< we added this
- old_key: NGINX_REQUEST # <<< we added this
-```
-
-Let's test it:
-
-```bash
-# echo '1.2.3.4 - - [19/Nov/2023:00:24:43 +0000] "GET /index.html HTTP/1.1" 200 4172 "-" "Go-http-client/1.1"' | log2journal -f nginx.yaml
-MESSAGE=GET /index.html HTTP/1.1 # <<< renamed !
-NGINX_BODY_BYTES_SENT=4172
-NGINX_HTTP_REFERER=-
-NGINX_HTTP_USER_AGENT=Go-http-client/1.1
-NGINX_REMOTE_ADDR=1.2.3.4
-NGINX_REMOTE_USER=-
-NGINX_REQUEST_METHOD=GET
-NGINX_REQUEST_URI=/index.html
-NGINX_SERVER_PROTOCOL=HTTP/1.1
-NGINX_STATUS=200
-NGINX_TIME_LOCAL=19/Nov/2023:00:24:43 +0000
-
-```
-
-### Injecting new fields
-
-To have a complete message in journals we need 3 fields: `MESSAGE`, `PRIORITY` and `SYSLOG_IDENTIFIER`. We have already added `MESSAGE` by renaming `NGINX_REQUEST`. We can also inject a `SYSLOG_IDENTIFIER` and `PRIORITY`.
-
-Ideally, we would want the 5xx errors to be red in our `journalctl` output and the dashboard. To achieve that we need to set the `PRIORITY` field to the right log level. Log priorities are numeric and follow the `syslog` priorities. Checking `/usr/include/sys/syslog.h` we can see these:
-
-```c
-#define LOG_EMERG 0 /* system is unusable */
-#define LOG_ALERT 1 /* action must be taken immediately */
-#define LOG_CRIT 2 /* critical conditions */
-#define LOG_ERR 3 /* error conditions */
-#define LOG_WARNING 4 /* warning conditions */
-#define LOG_NOTICE 5 /* normal but significant condition */
-#define LOG_INFO 6 /* informational */
-#define LOG_DEBUG 7 /* debug-level messages */
-```
-
-Avoid setting priority to 0 (`LOG_EMERG`), because these will be on your terminal (the journal uses `wall` to let you know of such events). A good priority for errors is 3 (red), or 4 (yellow).
-
-To set the PRIORITY field in the output, we can use `NGINX_STATUS`. We will do this in 2 steps: a) inject the priority field as a copy is `NGINX_STATUS` and then b) use a pattern on its value to rewrite it to the priority level we want.
-
-First, let's inject `SYSLOG_IDENTIFIER` and `PRIORITY`:
-
-```yaml
-pattern: |
- (?x) # Enable PCRE2 extended mode
- ^
- (?<remote_addr>[^ ]+) \s - \s
- (?<remote_user>[^ ]+) \s
- \[
- (?<time_local>[^\]]+)
- \]
- \s+ "
- (?<request>
- (?<request_method>[A-Z]+) \s+
- (?<request_uri>[^ ]+) \s+
- (?<server_protocol>[^"]+)
- )
- " \s+
- (?<status>\d+) \s+
- (?<body_bytes_sent>\d+) \s+
- "(?<http_referer>[^"]*)" \s+
- "(?<http_user_agent>[^"]*)"
-
-prefix: 'NGINX_'
-
-rename:
- - new_key: MESSAGE
- old_key: NGINX_REQUEST
-
-inject: # <<< we added this
- - key: PRIORITY # <<< we added this
- value: '${NGINX_STATUS}' # <<< we added this
-
- - key: SYSLOG_IDENTIFIER # <<< we added this
- value: 'nginx-log' # <<< we added this
-```
-
-Let's see what this does:
-
-```bash
-# echo '1.2.3.4 - - [19/Nov/2023:00:24:43 +0000] "GET /index.html HTTP/1.1" 200 4172 "-" "Go-http-client/1.1"' | log2journal -f nginx.yaml
-MESSAGE=GET /index.html HTTP/1.1
-NGINX_BODY_BYTES_SENT=4172
-NGINX_HTTP_REFERER=-
-NGINX_HTTP_USER_AGENT=Go-http-client/1.1
-NGINX_REMOTE_ADDR=1.2.3.4
-NGINX_REMOTE_USER=-
-NGINX_REQUEST_METHOD=GET
-NGINX_REQUEST_URI=/index.html
-NGINX_SERVER_PROTOCOL=HTTP/1.1
-NGINX_STATUS=200
-NGINX_TIME_LOCAL=19/Nov/2023:00:24:43 +0000
-PRIORITY=200 # <<< PRIORITY added
-SYSLOG_IDENTIFIER=nginx-log # <<< SYSLOG_IDENTIFIER added
-
-```
-
-### Rewriting field values
-
-Now we need to rewrite `PRIORITY` to the right syslog level based on its value (`NGINX_STATUS`). We will assign the priority 6 (info) when the status is 1xx, 2xx, 3xx, priority 5 (notice) when status is 4xx, priority 3 (error) when status is 5xx and anything else will go to priority 4 (warning). Let's do it:
-
-```yaml
-pattern: |
- (?x) # Enable PCRE2 extended mode
- ^
- (?<remote_addr>[^ ]+) \s - \s
- (?<remote_user>[^ ]+) \s
- \[
- (?<time_local>[^\]]+)
- \]
- \s+ "
- (?<request>
- (?<request_method>[A-Z]+) \s+
- (?<request_uri>[^ ]+) \s+
- (?<server_protocol>[^"]+)
- )
- " \s+
- (?<status>\d+) \s+
- (?<body_bytes_sent>\d+) \s+
- "(?<http_referer>[^"]*)" \s+
- "(?<http_user_agent>[^"]*)"
-
-prefix: 'NGINX_'
-
-rename:
- - new_key: MESSAGE
- old_key: NGINX_REQUEST
-
-inject:
- - key: PRIORITY
- value: '${NGINX_STATUS}'
-
-rewrite: # <<< we added this
- - key: PRIORITY # <<< we added this
- match: '^[123]' # <<< we added this
- value: 6 # <<< we added this
-
- - key: PRIORITY # <<< we added this
- match: '^4' # <<< we added this
- value: 5 # <<< we added this
-
- - key: PRIORITY # <<< we added this
- match: '^5' # <<< we added this
- value: 3 # <<< we added this
-
- - key: PRIORITY # <<< we added this
- match: '.*' # <<< we added this
- value: 4 # <<< we added this
-```
-
-Rewrite rules are processed in order and the first matching a field, stops by default processing for this field. This is why the last rule, that matches everything does not always change the priority to 4.
-
-Let's test it:
-
-```bash
-# echo '1.2.3.4 - - [19/Nov/2023:00:24:43 +0000] "GET /index.html HTTP/1.1" 200 4172 "-" "Go-http-client/1.1"' | log2journal -f nginx.yaml
-MESSAGE=GET /index.html HTTP/1.1
-NGINX_BODY_BYTES_SENT=4172
-NGINX_HTTP_REFERER=-
-NGINX_HTTP_USER_AGENT=Go-http-client/1.1
-NGINX_REMOTE_ADDR=1.2.3.4
-NGINX_REMOTE_USER=-
-NGINX_REQUEST_METHOD=GET
-NGINX_REQUEST_URI=/index.html
-NGINX_SERVER_PROTOCOL=HTTP/1.1
-NGINX_STATUS=200
-NGINX_TIME_LOCAL=19/Nov/2023:00:24:43 +0000
-PRIORITY=6 # <<< PRIORITY rewritten here
-SYSLOG_IDENTIFIER=nginx-log
-
-```
-
-Rewrite rules are powerful. You can have named groups in them, like in the main pattern, to extract sub-fields from them, which you can then use in variable substitution. You can use rewrite rules to anonymize the URLs, e.g to remove customer IDs or transaction details from them.
-
-### Sending logs to systemd-journal
-
-Now the message is ready to be sent to a systemd-journal. For this we use `systemd-cat-native`. This command can send such messages to a journal running on the localhost, a local journal namespace, or a `systemd-journal-remote` running on another server. By just appending `| systemd-cat-native` to the command, the message will be sent to the local journal.
-
-
-```bash
-# echo '1.2.3.4 - - [19/Nov/2023:00:24:43 +0000] "GET /index.html HTTP/1.1" 200 4172 "-" "Go-http-client/1.1"' | log2journal -f nginx.yaml | systemd-cat-native
-# no output
-
-# let's find the message
-# journalctl -r -o verbose SYSLOG_IDENTIFIER=nginx-log
-Wed 2023-12-06 13:23:07.083299 EET [s=5290f0133f25407aaa1e2c451c0e4756;i=57194;b=0dfa96ecc2094cecaa8ec0efcb93b865;m=b133308867;t=60bd59346a289;x=5c1bdacf2b9c4bbd]
- PRIORITY=6
- _UID=0
- _GID=0
- _CAP_EFFECTIVE=1ffffffffff
- _SELINUX_CONTEXT=unconfined
- _BOOT_ID=0dfa96ecc2094cecaa8ec0efcb93b865
- _MACHINE_ID=355c8eca894d462bbe4c9422caf7a8bb
- _HOSTNAME=lab-logtest-src
- _RUNTIME_SCOPE=system
- _TRANSPORT=journal
- MESSAGE=GET /index.html HTTP/1.1
- NGINX_BODY_BYTES_SENT=4172
- NGINX_HTTP_REFERER=-
- NGINX_HTTP_USER_AGENT=Go-http-client/1.1
- NGINX_REMOTE_ADDR=1.2.3.4
- NGINX_REMOTE_USER=-
- NGINX_REQUEST_METHOD=GET
- NGINX_REQUEST_URI=/index.html
- NGINX_SERVER_PROTOCOL=HTTP/1.1
- NGINX_STATUS=200
- NGINX_TIME_LOCAL=19/Nov/2023:00:24:43 +0000
- SYSLOG_IDENTIFIER=nginx-log
- _PID=114343
- _COMM=systemd-cat-nat
- _AUDIT_SESSION=253
- _AUDIT_LOGINUID=1000
- _SYSTEMD_CGROUP=/user.slice/user-1000.slice/session-253.scope
- _SYSTEMD_SESSION=253
- _SYSTEMD_OWNER_UID=1000
- _SYSTEMD_UNIT=session-253.scope
- _SYSTEMD_SLICE=user-1000.slice
- _SYSTEMD_USER_SLICE=-.slice
- _SYSTEMD_INVOCATION_ID=c59e33ead8c24880b027e317b89f9f76
- _SOURCE_REALTIME_TIMESTAMP=1701861787083299
-
-```
-
-So, the log line, with all its fields parsed, ended up in systemd-journal. Now we can send all the nginx logs to systemd-journal like this:
-
-```bash
-tail -F /var/log/nginx/access.log |\
- log2journal -f nginx.yaml |\
- systemd-cat-native
-```
-
-## Best practices
-
-**Create a systemd service unit**: Add the above commands to a systemd unit file. When you run it in a systemd unit file you will be able to start/stop it and also see its status. Furthermore you can use the `LogNamespace=` directive of systemd service units to isolate your nginx logs from the logs of the rest of the system. Here is how to do it:
-
-Create the file `/etc/systemd/system/nginx-logs.service` (change `/path/to/nginx.yaml` to the right path):
-
-```
-[Unit]
-Description=NGINX Log to Systemd Journal
-After=network.target
-
-[Service]
-ExecStart=/bin/sh -c 'tail -F /var/log/nginx/access.log | log2journal -f /path/to/nginx.yaml' | systemd-cat-native
-LogNamespace=nginx-logs
-Restart=always
-RestartSec=3
-
-[Install]
-WantedBy=multi-user.target
-```
-
-Reload systemd to grab this file:
-
-```bash
-sudo systemctl daemon-reload
-```
-
-Enable and start the service:
-
-```bash
-sudo systemctl enable nginx-logs.service
-sudo systemctl start nginx-logs.service
-```
-
-To see the logs of the namespace, use:
-
-```bash
-journalctl -f --namespace=nginx-logs
-```
-
-Netdata will automatically pick the new namespace and present it at the list of sources of the dashboard.
-
-You can also instruct `systemd-cat-native` to log to a remote system, sending the logs to a `systemd-journal-remote` instance running on another server. Check [the manual of systemd-cat-native](https://github.com/netdata/netdata/blob/master/libnetdata/log/systemd-cat-native.md).
-
-
-## Performance
-
-`log2journal` and `systemd-cat-native` have been designed to process hundreds of thousands of log lines per second. They both utilize high performance indexing hashtables to speed up lookups, and queues that dynamically adapt to the number of log lines offered, offering a smooth and fast experience under all conditions.
-
-In our tests, the combined CPU utilization of `log2journal` and `systemd-cat-native` versus `promtail` with similar configuration is 1 to 5. So, `log2journal` and `systemd-cat-native` combined, are 5 times faster than `promtail`.
-
-### PCRE2 patterns
-
-The key characteristic that can influence the performance of a logs processing pipeline using these tools, is the quality of the PCRE2 patterns used. Poorly created PCRE2 patterns can make processing significantly slower, or CPU consuming.
-
-Especially the pattern `.*` seems to have the biggest impact on CPU consumption, especially when multiple `.*` are on the same pattern.
-
-Usually we use `.*` to indicate that we need to match everything up to a character, e.g. `.* ` to match up to a space. By replacing it with `[^ ]+` (meaning: match at least a character up to a space), the regular expression engine can be a lot more efficient, reducing the overall CPU utilization significantly.
-
-### Performance of systemd journals
-
-The ingestion pipeline of logs, from `tail` to `systemd-journald` or `systemd-journal-remote` is very efficient in all aspects. CPU utilization is better than any other system we tested and RAM usage is independent of the number of fields indexed, making systemd-journal one of the most efficient log management engines for ingesting high volumes of structured logs.
-
-High fields cardinality does not have a noticable impact on systemd-journal. The amount of fields indexed and the amount of unique values per field, have a linear and predictable result in the resource utilization of `systemd-journald` and `systemd-journal-remote`. This is unlike other logs management solutions, like Loki, that their RAM requirements grow exponentially as the cardinality increases, making it impractical for them to index the amount of information systemd journals can index.
-
-However, the number of fields added to journals influences the overall disk footprint. Less fields means more log entries per journal file, smaller overall disk footprint and faster queries.
-
-systemd-journal files are primarily designed for security and reliability. This comes at the cost of disk footprint. The internal structure of journal files is such that in case of corruption, minimum data loss will incur. To achieve such a unique characteristic, certain data within the files need to be aligned at predefined boundaries, so that in case there is a corruption, non-corrupted parts of the journal file can be recovered.
-
-Despite the fact that systemd-journald employees several techniques to optimize disk footprint, like deduplication of log entries, shared indexes for fields and their values, compression of long log entries, etc. the disk footprint of journal files is generally 10x more compared to other monitoring solutions, like Loki.
-
-This can be improved by storing journal files in a compressed filesystem. In our tests, a compressed filesystem can save up to 75% of the space required by journal files. The journal files will still be bigger than the overall disk footprint of other solutions, but the flexibility (index any number of fields), reliability (minimal potential data loss) and security (tampering protection and sealing) features of systemd-journal justify the difference.
-
-When using versions of systemd prior to 254 and you are centralizing logs to a remote system, `systemd-journal-remote` creates very small files (32MB). This results in increased duplication of information across the files, increasing the overall disk footprint. systemd versions 254+, added options to `systemd-journal-remote` to control the max size per file. This can significantly reduce the duplication of information.
-
-Another limitation of the `systemd-journald` ecosystem is the uncompressed transmission of logs across systems. `systemd-journal-remote` up to version 254 that we tested, accepts encrypted, but uncompressed data. This means that when centralizing logs to a logs server, the bandwidth required will be increased compared to other log management solution.
-
-## Security Considerations
-
-`log2journal` and `systemd-cat-native` are used to convert log files to structured logs in the systemd-journald ecosystem.
-
-Systemd-journal is a logs management solution designed primarily for security and reliability. When configured properly, it can reliably and securely store your logs, ensuring they will available and unchanged for as long as you need them.
-
-When sending logs to a remote system, `systemd-cat-native` can be configured the same way `systemd-journal-upload` is configured, using HTTPS and private keys to encrypt and secure their transmission over the network.
-
-When dealing with sensitive logs, organizations usually follow 2 strategies:
-
-1. Anonymize the logs before storing them, so that the stored logs do not have any sensitive information.
-2. Store the logs in full, including sensitive information, and carefully control who and how has access to them.
-
-Netdata can help in both cases.
-
-If you want to anonymize the logs before storing them, use rewriting rules at the `log2journal` phase to remove sensitive information from them. This process usually means matching the sensitive part and replacing with `XXX` or `CUSTOMER_ID`, or `CREDIT_CARD_NUMBER`, so that the resulting log entries stored in journal files will not include any such sensitive information.
-
-If on other hand your organization prefers to maintain the full logs and control who and how has access on them, use Netdata Cloud to assign roles to your team members and control which roles can access the journal logs in your environment.
-
-## `log2journal` options
-
-```
-
-Netdata log2journal v1.43.0-341-gdac4df856
-
-Convert logs to systemd Journal Export Format.
-
- - JSON logs: extracts all JSON fields.
- - logfmt logs: extracts all logfmt fields.
- - free-form logs: uses PCRE2 patterns to extracts fields.
-
-Usage: ./log2journal [OPTIONS] PATTERN|json
-
-Options:
-
- --file /path/to/file.yaml or -f /path/to/file.yaml
- Read yaml configuration file for instructions.
-
- --config CONFIG_NAME or -c CONFIG_NAME
- Run with the internal YAML configuration named CONFIG_NAME.
- Available internal YAML configs:
-
- nginx-combined nginx-json default
-
---------------------------------------------------------------------------------
- INPUT PROCESSING
-
- PATTERN
- PATTERN should be a valid PCRE2 regular expression.
- RE2 regular expressions (like the ones usually used in Go applications),
- are usually valid PCRE2 patterns too.
- Sub-expressions without named groups are evaluated, but their matches are
- not added to the output.
-
- - JSON mode
- JSON mode is enabled when the pattern is set to: json
- Field names are extracted from the JSON logs and are converted to the
- format expected by Journal Export Format (all caps, only _ is allowed).
-
- - logfmt mode
- logfmt mode is enabled when the pattern is set to: logfmt
- Field names are extracted from the logfmt logs and are converted to the
- format expected by Journal Export Format (all caps, only _ is allowed).
-
- All keys extracted from the input, are transliterated to match Journal
- semantics (capital A-Z, digits 0-9, underscore).
-
- In a YAML file:
- ```yaml
- pattern: 'PCRE2 pattern | json | logfmt'
- ```
-
---------------------------------------------------------------------------------
- GLOBALS
-
- --prefix PREFIX
- Prefix all fields with PREFIX. The PREFIX is added before any other
- processing, so that the extracted keys have to be matched with the PREFIX in
- them. PREFIX is NOT transliterated and it is assumed to be systemd-journal
- friendly.
-
- In a YAML file:
- ```yaml
- prefix: 'PREFIX_' # prepend all keys with this prefix.
- ```
-
- --filename-key KEY
- Add a field with KEY as the key and the current filename as value.
- Automatically detects filenames when piped after 'tail -F',
- and tail matches multiple filenames.
- To inject the filename when tailing a single file, use --inject.
-
- In a YAML file:
- ```yaml
- filename:
- key: KEY
- ```
-
---------------------------------------------------------------------------------
- RENAMING OF KEYS
-
- --rename NEW=OLD
- Rename fields. OLD has been transliterated and PREFIX has been added.
- NEW is assumed to be systemd journal friendly.
-
- Up to 512 renaming rules are allowed.
-
- In a YAML file:
- ```yaml
- rename:
- - new_key: KEY1
- old_key: KEY2 # transliterated with PREFIX added
- - new_key: KEY3
- old_key: KEY4 # transliterated with PREFIX added
- # add as many as required
- ```
-
---------------------------------------------------------------------------------
- INJECTING NEW KEYS
-
- --inject KEY=VALUE
- Inject constant fields to the output (both matched and unmatched logs).
- --inject entries are added to unmatched lines too, when their key is
- not used in --inject-unmatched (--inject-unmatched override --inject).
- VALUE can use variable like ${OTHER_KEY} to be replaced with the values
- of other keys available.
-
- Up to 512 fields can be injected.
-
- In a YAML file:
- ```yaml
- inject:
- - key: KEY1
- value: 'VALUE1'
- - key: KEY2
- value: '${KEY3}${KEY4}' # gets the values of KEY3 and KEY4
- # add as many as required
- ```
-
---------------------------------------------------------------------------------
- REWRITING KEY VALUES
-
- --rewrite KEY=/MATCH/REPLACE[/OPTIONS]
- Apply a rewrite rule to the values of a specific key.
- The first character after KEY= is the separator, which should also
- be used between the MATCH, REPLACE and OPTIONS.
-
- OPTIONS can be a comma separated list of `non-empty`, `dont-stop` and
- `inject`.
-
- When `non-empty` is given, MATCH is expected to be a variable
- substitution using `${KEY1}${KEY2}`. Once the substitution is completed
- the rule is matching the KEY only if the result is not empty.
- When `non-empty` is not set, the MATCH string is expected to be a PCRE2
- regular expression to be checked against the KEY value. This PCRE2
- pattern may include named groups to extract parts of the KEY's value.
-
- REPLACE supports variable substitution like `${variable}` against MATCH
- named groups (when MATCH is a PCRE2 pattern) and `${KEY}` against the
- keys defined so far.
-
- Example:
- --rewrite DATE=/^(?<year>\d{4})-(?<month>\d{2})-(?<day>\d{2})$/
- ${day}/${month}/${year}
- The above will rewrite dates in the format YYYY-MM-DD to DD/MM/YYYY.
-
- Only one rewrite rule is applied per key; the sequence of rewrites for a
- given key, stops once a rule matches it. This allows providing a sequence
- of independent rewriting rules for the same key, matching the different
- values the key may get, and also provide a catch-all rewrite rule at the
- end, for setting the key value if no other rule matched it. The rewrite
- rule can allow processing more rewrite rules when OPTIONS includes
- the keyword 'dont-stop'.
-
- Up to 512 rewriting rules are allowed.
-
- In a YAML file:
- ```yaml
- rewrite:
- # the order if these rules in important - processed top to bottom
- - key: KEY1
- match: 'PCRE2 PATTERN WITH NAMED GROUPS'
- value: 'all match fields and input keys as ${VARIABLE}'
- inject: BOOLEAN # yes = inject the field, don't just rewrite it
- stop: BOOLEAN # no = continue processing, don't stop if matched
- - key: KEY2
- non_empty: '${KEY3}${KEY4}' # match only if this evaluates to non empty
- value: 'all input keys as ${VARIABLE}'
- inject: BOOLEAN # yes = inject the field, don't just rewrite it
- stop: BOOLEAN # no = continue processing, don't stop if matched
- # add as many rewrites as required
- ```
-
- By default rewrite rules are applied only on fields already defined.
- This allows shipping YAML files that include more rewrites than are
- required for a specific input file.
- Rewrite rules however allow injecting new fields when OPTIONS include
- the keyword `inject` or in YAML `inject: yes` is given.
-
- MATCH on the command line can be empty to define an unconditional rule.
- Similarly, `match` and `non_empty` can be omitted in the YAML file.
---------------------------------------------------------------------------------
- UNMATCHED LINES
-
- --unmatched-key KEY
- Include unmatched log entries in the output with KEY as the field name.
- Use this to include unmatched entries to the output stream.
- Usually it should be set to --unmatched-key=MESSAGE so that the
- unmatched entry will appear as the log message in the journals.
- Use --inject-unmatched to inject additional fields to unmatched lines.
-
- In a YAML file:
- ```yaml
- unmatched:
- key: MESSAGE # inject the error log as MESSAGE
- ```
-
- --inject-unmatched LINE
- Inject lines into the output for each unmatched log entry.
- Usually, --inject-unmatched=PRIORITY=3 is needed to mark the unmatched
- lines as errors, so that they can easily be spotted in the journals.
-
- Up to 512 such lines can be injected.
-
- In a YAML file:
- ```yaml
- unmatched:
- key: MESSAGE # inject the error log as MESSAGE
- inject::
- - key: KEY1
- value: 'VALUE1'
- # add as many constants as required
- ```
-
---------------------------------------------------------------------------------
- FILTERING
-
- --include PATTERN
- Include only keys matching the PCRE2 PATTERN.
- Useful when parsing JSON of logfmt logs, to include only the keys given.
- The keys are matched after the PREFIX has been added to them.
-
- --exclude PATTERN
- Exclude the keys matching the PCRE2 PATTERN.
- Useful when parsing JSON of logfmt logs, to exclude some of the keys given.
- The keys are matched after the PREFIX has been added to them.
-
- When both include and exclude patterns are set and both match a key,
- exclude wins and the key will not be added, like a pipeline, we first
- include it and then exclude it.
-
- In a YAML file:
- ```yaml
- filter:
- include: 'PCRE2 PATTERN MATCHING KEY NAMES TO INCLUDE'
- exclude: 'PCRE2 PATTERN MATCHING KEY NAMES TO EXCLUDE'
- ```
-
---------------------------------------------------------------------------------
- OTHER
-
- -h, or --help
- Display this help and exit.
-
- --show-config
- Show the configuration in YAML format before starting the job.
- This is also an easy way to convert command line parameters to yaml.
-
-The program accepts all parameters as both --option=value and --option value.
-
-The maximum log line length accepted is 1048576 characters.
-
-PIPELINE AND SEQUENCE OF PROCESSING
-
-This is a simple diagram of the pipeline taking place:
-
- +---------------------------------------------------+
- | INPUT |
- | read one log line at a time |
- +---------------------------------------------------+
- v v v v v v
- +---------------------------------------------------+
- | EXTRACT FIELDS AND VALUES |
- | JSON, logfmt, or pattern based |
- | (apply optional PREFIX - all keys use capitals) |
- +---------------------------------------------------+
- v v v v v v
- +---------------------------------------------------+
- | RENAME FIELDS |
- | change the names of the fields |
- +---------------------------------------------------+
- v v v v v v
- +---------------------------------------------------+
- | INJECT NEW FIELDS |
- | constants, or other field values as variables |
- +---------------------------------------------------+
- v v v v v v
- +---------------------------------------------------+
- | REWRITE FIELD VALUES |
- | pipeline multiple rewriting rules to alter |
- | the values of the fields |
- +---------------------------------------------------+
- v v v v v v
- +---------------------------------------------------+
- | FILTER FIELDS |
- | use include and exclude patterns on the field |
- | names, to select which fields are sent to journal |
- +---------------------------------------------------+
- v v v v v v
- +---------------------------------------------------+
- | OUTPUT |
- | generate Journal Export Format |
- +---------------------------------------------------+
-
---------------------------------------------------------------------------------
-JOURNAL FIELDS RULES (enforced by systemd-journald)
-
- - field names can be up to 64 characters
- - the only allowed field characters are A-Z, 0-9 and underscore
- - the first character of fields cannot be a digit
- - protected journal fields start with underscore:
- * they are accepted by systemd-journal-remote
- * they are NOT accepted by a local systemd-journald
-
- For best results, always include these fields:
-
- MESSAGE=TEXT
- The MESSAGE is the body of the log entry.
- This field is what we usually see in our logs.
-
- PRIORITY=NUMBER
- PRIORITY sets the severity of the log entry.
- 0=emerg, 1=alert, 2=crit, 3=err, 4=warn, 5=notice, 6=info, 7=debug
- - Emergency events (0) are usually broadcast to all terminals.
- - Emergency, alert, critical, and error (0-3) are usually colored red.
- - Warning (4) entries are usually colored yellow.
- - Notice (5) entries are usually bold or have a brighter white color.
- - Info (6) entries are the default.
- - Debug (7) entries are usually grayed or dimmed.
-
- SYSLOG_IDENTIFIER=NAME
- SYSLOG_IDENTIFIER sets the name of application.
- Use something descriptive, like: SYSLOG_IDENTIFIER=nginx-logs
-
-You can find the most common fields at 'man systemd.journal-fields'.
-
-```
-
-`log2journal` supports YAML configuration files, like the ones found [in this directory](https://github.com/netdata/netdata/tree/master/collectors/log2journal/log2journal.d).
-
-## `systemd-cat-native` options
-
-Read [the manual of systemd-cat-native](https://github.com/netdata/netdata/blob/master/libnetdata/log/systemd-cat-native.md).
diff --git a/collectors/macos.plugin/Makefile.am b/collectors/macos.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/macos.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/macos.plugin/metadata.yaml b/collectors/macos.plugin/metadata.yaml
deleted file mode 100644
index cc159ad1f..000000000
--- a/collectors/macos.plugin/metadata.yaml
+++ /dev/null
@@ -1,727 +0,0 @@
-plugin_name: macos.plugin
-modules:
- - meta:
- plugin_name: macos.plugin
- module_name: mach_smi
- monitored_instance:
- name: macOS
- link: "https://www.apple.com/macos"
- categories:
- - data-collection.macos-systems
- icon_filename: "macos.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - macos
- - apple
- - darwin
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor macOS metrics for efficient operating system performance."
- method_description: |
- The plugin uses three different methods to collect data:
- - The function `sysctlbyname` is called to collect network, swap, loadavg, and boot time.
- - The functtion `host_statistic` is called to collect CPU and Virtual memory data;
- - The function `IOServiceGetMatchingServices` to collect storage information.
- supported_platforms:
- include:
- - macOS
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- description: "The netdata main configuration file."
- options:
- description: |
- There are three sections in the file which you can configure:
-
- - `[plugin:macos:sysctl]` - Enable or disable monitoring for network, swap, loadavg, and boot time.
- - `[plugin:macos:mach_smi]` - Enable or disable monitoring for CPU and Virtual memory.
- - `[plugin:macos:iokit]` - Enable or disable monitoring for storage device.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: enable load average
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of load average metrics (load1, load5, load15).
- default_value: yes
- required: false
- - name: system swap
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of system swap metrics (free, used).
- default_value: yes
- required: false
- - name: bandwidth
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of network bandwidth metrics (received, sent).
- default_value: yes
- required: false
- - name: ipv4 TCP packets
- description: Enable or disable monitoring of IPv4 TCP total packets metrics (received, sent).
- section_name: plugin:macos:sysctl
- default_value: yes
- required: false
- - name: ipv4 TCP errors
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of IPv4 TCP packets metrics (Input Errors, Checksum, Retransmission segments).
- default_value: yes
- required: false
- - name: ipv4 TCP handshake issues
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of IPv4 TCP handshake metrics (Established Resets, Active Opens, Passive Opens, Attempt Fails).
- default_value: yes
- required: false
- - name: ECN packets
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ECN statistics metrics (InCEPkts, InNoECTPkts).
- default_value: auto
- required: false
- - name: TCP SYN cookies
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of TCP SYN cookies metrics (received, sent, failed).
- default_value: auto
- required: false
- - name: TCP out-of-order queue
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of TCP out-of-order queue metrics (inqueue).
- default_value: auto
- required: false
- - name: TCP connection aborts
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of TCP connection aborts metrics (Bad Data, User closed, No memory, Timeout).
- default_value: auto
- required: false
- - name: ipv4 UDP packets
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ipv4 UDP packets metrics (sent, received.).
- default_value: yes
- required: false
- - name: ipv4 UDP errors
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ipv4 UDP errors metrics (Recieved Buffer error, Input Errors, No Ports, IN Checksum Errors, Ignore Multi).
- default_value: yes
- required: false
- - name: ipv4 icmp packets
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of IPv4 ICMP packets metrics (sent, received, in error, OUT error, IN Checksum error).
- default_value: yes
- required: false
- - name: ipv4 icmp messages
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ipv4 ICMP messages metrics (I/O messages, I/O Errors, In Checksum).
- default_value: yes
- required: false
- - name: ipv4 packets
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ipv4 packets metrics (received, sent, forwarded, delivered).
- default_value: yes
- required: false
- - name: ipv4 fragments sent
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of IPv4 fragments sent metrics (ok, fails, creates).
- default_value: yes
- required: false
- - name: ipv4 fragments assembly
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of IPv4 fragments assembly metrics (ok, failed, all).
- default_value: yes
- required: false
- - name: ipv4 errors
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of IPv4 errors metrics (I/O discard, I/O HDR errors, In Addr errors, In Unknown protos, OUT No Routes).
- default_value: yes
- required: false
- - name: ipv6 packets
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of IPv6 packets metrics (received, sent, forwarded, delivered).
- default_value: auto
- required: false
- - name: ipv6 fragments sent
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of IPv6 fragments sent metrics (ok, failed, all).
- default_value: auto
- required: false
- - name: ipv6 fragments assembly
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of IPv6 fragments assembly metrics (ok, failed, timeout, all).
- default_value: auto
- required: false
- - name: ipv6 errors
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of IPv6 errors metrics (I/O Discards, In Hdr Errors, In Addr Errors, In Truncaedd Packets, I/O No Routes).
- default_value: auto
- required: false
- - name: icmp
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ICMP metrics (sent, received).
- default_value: auto
- required: false
- - name: icmp redirects
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ICMP redirects metrics (received, sent).
- default_value: auto
- required: false
- - name: icmp errors
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ICMP metrics (I/O Errors, In Checksums, In Destination Unreachable, In Packet too big, In Time Exceeds, In Parm Problem, Out Dest Unreachable, Out Timee Exceeds, Out Parm Problems.).
- default_value: auto
- required: false
- - name: icmp echos
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ICMP echos metrics (I/O Echos, I/O Echo Reply).
- default_value: auto
- required: false
- - name: icmp router
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ICMP router metrics (I/O Solicits, I/O Advertisements).
- default_value: auto
- required: false
- - name: icmp neighbor
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ICMP neighbor metrics (I/O Solicits, I/O Advertisements).
- default_value: auto
- required: false
- - name: icmp types
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of ICMP types metrics (I/O Type1, I/O Type128, I/O Type129, Out Type133, Out Type135, In Type136, Out Type145).
- default_value: auto
- required: false
- - name: space usage for all disks
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of space usage for all disks metrics (available, used, reserved for root).
- default_value: yes
- required: false
- - name: inodes usage for all disks
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of inodes usage for all disks metrics (available, used, reserved for root).
- default_value: yes
- required: false
- - name: bandwidth
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of bandwidth metrics (received, sent).
- default_value: yes
- required: false
- - name: system uptime
- section_name: plugin:macos:sysctl
- description: Enable or disable monitoring of system uptime metrics (uptime).
- default_value: yes
- required: false
- - name: cpu utilization
- section_name: plugin:macos:mach_smi
- description: Enable or disable monitoring of CPU utilization metrics (user, nice, system, idel).
- default_value: yes
- required: false
- - name: system ram
- section_name: plugin:macos:mach_smi
- description: Enable or disable monitoring of system RAM metrics (Active, Wired, throttled, compressor, inactive, purgeable, speculative, free).
- default_value: yes
- required: false
- - name: swap i/o
- section_name: plugin:macos:mach_smi
- description: Enable or disable monitoring of SWAP I/O metrics (I/O Swap).
- default_value: yes
- required: false
- - name: memory page faults
- section_name: plugin:macos:mach_smi
- description: Enable or disable monitoring of memory page faults metrics (memory, cow, I/O page, compress, decompress, zero fill, reactivate, purge).
- default_value: yes
- required: false
- - name: disk i/o
- section_name: plugin:macos:iokit
- description: Enable or disable monitoring of disk I/O metrics (In, Out).
- default_value: yes
- required: false
- examples:
- folding:
- enabled: false
- title: "Config"
- list:
- - name: Disable swap monitoring.
- folding:
- enabled: true
- description: A basic example that discards swap monitoring
- config: |
- [plugin:macos:sysctl]
- system swap = no
- [plugin:macos:mach_smi]
- swap i/o = no
- - name: Disable complete Machine SMI section.
- folding:
- enabled: true
- description: A basic example that discards swap monitoring
- config: |
- [plugin:macos:mach_smi]
- cpu utilization = no
- system ram = no
- swap i/o = no
- memory page faults = no
- disk i/o = no
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: interface_speed
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.net
- info: network interface ${label:device} current speed
- os: "*"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: |
- These metrics refer to hardware and network monitoring.
- labels: []
- metrics:
- - name: system.cpu
- description: Total CPU utilization
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: user
- - name: nice
- - name: system
- - name: idle
- - name: system.ram
- description: System RAM
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: active
- - name: wired
- - name: throttled
- - name: compressor
- - name: inactive
- - name: purgeable
- - name: speculative
- - name: free
- - name: mem.swapio
- description: Swap I/O
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: io
- - name: out
- - name: mem.pgfaults
- description: Memory Page Faults
- unit: "faults/s"
- chart_type: line
- dimensions:
- - name: memory
- - name: cow
- - name: pagein
- - name: pageout
- - name: compress
- - name: decompress
- - name: zero_fill
- - name: reactivate
- - name: purge
- - name: system.load
- description: System Load Average
- unit: "load"
- chart_type: line
- dimensions:
- - name: load1
- - name: load5
- - name: load15
- - name: mem.swap
- description: System Swap
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: free
- - name: used
- - name: system.ipv4
- description: IPv4 Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.tcppackets
- description: IPv4 TCP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.tcperrors
- description: IPv4 TCP Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InErrs
- - name: InCsumErrors
- - name: RetransSegs
- - name: ipv4.tcphandshake
- description: IPv4 TCP Handshake Issues
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: EstabResets
- - name: ActiveOpens
- - name: PassiveOpens
- - name: AttemptFails
- - name: ipv4.tcpconnaborts
- description: TCP Connection Aborts
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: baddata
- - name: userclosed
- - name: nomemory
- - name: timeout
- - name: ipv4.tcpofo
- description: TCP Out-Of-Order Queue
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: inqueue
- - name: ipv4.tcpsyncookies
- description: TCP SYN Cookies
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: failed
- - name: ipv4.ecnpkts
- description: IPv4 ECN Statistics
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: CEP
- - name: NoECTP
- - name: ipv4.udppackets
- description: IPv4 UDP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.udperrors
- description: IPv4 UDP Errors
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: RcvbufErrors
- - name: InErrors
- - name: NoPorts
- - name: InCsumErrors
- - name: IgnoredMulti
- - name: ipv4.icmp
- description: IPv4 ICMP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.icmp_errors
- description: IPv4 ICMP Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InErrors
- - name: OutErrors
- - name: InCsumErrors
- - name: ipv4.icmpmsg
- description: IPv4 ICMP Messages
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InEchoReps
- - name: OutEchoReps
- - name: InEchos
- - name: OutEchos
- - name: ipv4.packets
- description: IPv4 Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: forwarded
- - name: delivered
- - name: ipv4.fragsout
- description: IPv4 Fragments Sent
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: created
- - name: ipv4.fragsin
- description: IPv4 Fragments Reassembly
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: all
- - name: ipv4.errors
- description: IPv4 Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InDiscards
- - name: OutDiscards
- - name: InHdrErrors
- - name: OutNoRoutes
- - name: InAddrErrors
- - name: InUnknownProtos
- - name: ipv6.packets
- description: IPv6 Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: forwarded
- - name: delivers
- - name: ipv6.fragsout
- description: IPv6 Fragments Sent
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: all
- - name: ipv6.fragsin
- description: IPv6 Fragments Reassembly
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: timeout
- - name: all
- - name: ipv6.errors
- description: IPv6 Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InDiscards
- - name: OutDiscards
- - name: InHdrErrors
- - name: InAddrErrors
- - name: InTruncatedPkts
- - name: InNoRoutes
- - name: OutNoRoutes
- - name: ipv6.icmp
- description: IPv6 ICMP Messages
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.icmpredir
- description: IPv6 ICMP Redirects
- unit: "redirects/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.icmperrors
- description: IPv6 ICMP Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: InErrors
- - name: OutErrors
- - name: InCsumErrors
- - name: InDestUnreachs
- - name: InPktTooBigs
- - name: InTimeExcds
- - name: InParmProblems
- - name: OutDestUnreachs
- - name: OutTimeExcds
- - name: OutParmProblems
- - name: ipv6.icmpechos
- description: IPv6 ICMP Echo
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InEchos
- - name: OutEchos
- - name: InEchoReplies
- - name: OutEchoReplies
- - name: ipv6.icmprouter
- description: IPv6 Router Messages
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InSolicits
- - name: OutSolicits
- - name: InAdvertisements
- - name: OutAdvertisements
- - name: ipv6.icmpneighbor
- description: IPv6 Neighbor Messages
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InSolicits
- - name: OutSolicits
- - name: InAdvertisements
- - name: OutAdvertisements
- - name: ipv6.icmptypes
- description: IPv6 ICMP Types
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InType1
- - name: InType128
- - name: InType129
- - name: InType136
- - name: OutType1
- - name: OutType128
- - name: OutType129
- - name: OutType133
- - name: OutType135
- - name: OutType143
- - name: system.uptime
- description: System Uptime
- unit: "seconds"
- chart_type: line
- dimensions:
- - name: uptime
- - name: system.io
- description: Disk I/O
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: in
- - name: out
- - name: disk
- description: ""
- labels: []
- metrics:
- - name: disk.io
- description: Disk I/O Bandwidth
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: read
- - name: writes
- - name: disk.ops
- description: Disk Completed I/O Operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: read
- - name: writes
- - name: disk.util
- description: Disk Utilization Time
- unit: "% of time working"
- chart_type: area
- dimensions:
- - name: utilization
- - name: disk.iotime
- description: Disk Total I/O Time
- unit: "milliseconds/s"
- chart_type: line
- dimensions:
- - name: reads
- - name: writes
- - name: disk.await
- description: Average Completed I/O Operation Time
- unit: "milliseconds/operation"
- chart_type: line
- dimensions:
- - name: reads
- - name: writes
- - name: disk.avgsz
- description: Average Completed I/O Operation Bandwidth
- unit: "KiB/operation"
- chart_type: line
- dimensions:
- - name: reads
- - name: writes
- - name: disk.svctm
- description: Average Service Time
- unit: "milliseconds/operation"
- chart_type: line
- dimensions:
- - name: svctm
- - name: mount point
- description: ""
- labels: []
- metrics:
- - name: disk.space
- description: Disk Space Usage for {mounted dir} [{mounted filesystem}]
- unit: "GiB"
- chart_type: stacked
- dimensions:
- - name: avail
- - name: used
- - name: reserved_for_root
- - name: disk.inodes
- description: Disk Files (inodes) Usage for {mounted dir} [{mounted filesystem}]
- unit: "inodes"
- chart_type: stacked
- dimensions:
- - name: avail
- - name: used
- - name: reserved_for_root
- - name: network device
- description: ""
- labels: []
- metrics:
- - name: net.net
- description: Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: net.packets
- description: Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: multicast_received
- - name: multicast_sent
- - name: net.errors
- description: Interface Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: inbound
- - name: outbound
- - name: net.drops
- description: Interface Drops
- unit: "drops/s"
- chart_type: line
- dimensions:
- - name: inbound
- - name: net.events
- description: Network Interface Events
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: frames
- - name: collisions
- - name: carrier
diff --git a/collectors/nfacct.plugin/Makefile.am b/collectors/nfacct.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/nfacct.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/perf.plugin/Makefile.am b/collectors/perf.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/perf.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/perf.plugin/metadata.yaml b/collectors/perf.plugin/metadata.yaml
deleted file mode 100644
index eada3351d..000000000
--- a/collectors/perf.plugin/metadata.yaml
+++ /dev/null
@@ -1,252 +0,0 @@
-plugin_name: perf.plugin
-modules:
- - meta:
- plugin_name: perf.plugin
- module_name: perf.plugin
- monitored_instance:
- name: CPU performance
- link: "https://kernel.org/"
- categories:
- - data-collection.linux-systems
- icon_filename: "bolt.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - linux
- - cpu performance
- - cpu cache
- - perf.plugin
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This collector monitors CPU performance metrics about cycles, instructions, migrations, cache operations and more."
- method_description: "It uses syscall (2) to open a file descriptior to monitor the perf events."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: "It needs setuid to use necessary syscall to collect perf events. Netada sets the permission during installation time."
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: Install perf plugin
- description: |
- If you are [using our official native DEB/RPM packages](https://github.com/netdata/netdata/blob/master/packaging/installer/UPDATE.md#determine-which-installation-method-you-used), make sure the `netdata-plugin-perf` package is installed.
- - title: Enable the pref plugin
- description: |
- The plugin is disabled by default because the number of PMUs is usually quite limited and it is not desired to allow Netdata to struggle silently for PMUs, interfering with other performance monitoring software.
-
- To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`, to edit the `netdata.conf` file.
-
- ```bash
- cd /etc/netdata # Replace this path with your Netdata config directory, if different
- sudo ./edit-config netdata.conf
- ```
-
- Change the value of the `perf` setting to `yes` in the `[plugins]` section. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:perf]"
- description: "The main netdata configuration file."
- options:
- description: |
- You can get the available options running:
-
- ```bash
- /usr/libexec/netdata/plugins.d/perf.plugin --help
- ````
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 1
- required: false
- - name: command options
- description: Command options that specify charts shown by plugin. `cycles`, `instructions`, `branch`, `cache`, `bus`, `stalled`, `migrations`, `alignment`, `emulation`, `L1D`, `L1D-prefetch`, `L1I`, `LL`, `DTLB`, `ITLB`, `PBU`.
- default_value: 1
- required: true
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: All metrics
- folding:
- enabled: false
- description: Monitor all metrics available.
- config: |
- [plugin:perf]
- command options = all
- - name: CPU cycles
- description: Monitor CPU cycles.
- config: |
- [plugin:perf]
- command options = cycles
- troubleshooting:
- problems:
- list:
- - name: Debug Mode
- description: |
- You can run `perf.plugin` with the debug option enabled, to troubleshoot issues with it. The output should give you clues as to why the collector isn't working.
-
- - Navigate to the `plugins.d` directory, usually at `/usr/libexec/netdata/plugins.d/`. If that's not the case on your system, open `netdata.conf` and look for the `plugins` setting under `[directories]`.
-
- ```bash
- cd /usr/libexec/netdata/plugins.d/
- ```
-
- - Switch to the `netdata` user.
-
- ```bash
- sudo -u netdata -s
- ```
-
- - Run the `perf.plugin` in debug mode:
-
- ```bash
- ./perf.plugin 1 all debug
- ```
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: perf.cpu_cycles
- description: CPU cycles
- unit: "cycles/s"
- chart_type: line
- dimensions:
- - name: cpu
- - name: ref_cpu
- - name: perf.instructions
- description: Instructions
- unit: "instructions/s"
- chart_type: line
- dimensions:
- - name: instructions
- - name: perf.instructions_per_cycle
- description: Instructions per Cycle(IPC)
- unit: "instructions/cycle"
- chart_type: line
- dimensions:
- - name: ipc
- - name: perf.branch_instructions
- description: Branch instructions
- unit: "instructions/s"
- chart_type: line
- dimensions:
- - name: instructions
- - name: misses
- - name: perf.cache
- description: Cache operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: references
- - name: misses
- - name: perf.bus_cycles
- description: Bus cycles
- unit: "cycles/s"
- chart_type: line
- dimensions:
- - name: bus
- - name: perf.stalled_cycles
- description: Stalled frontend and backend cycles
- unit: "cycles/s"
- chart_type: line
- dimensions:
- - name: frontend
- - name: backend
- - name: perf.migrations
- description: CPU migrations
- unit: "migrations"
- chart_type: line
- dimensions:
- - name: migrations
- - name: perf.alignment_faults
- description: Alignment faults
- unit: "faults"
- chart_type: line
- dimensions:
- - name: faults
- - name: perf.emulation_faults
- description: Emulation faults
- unit: "faults"
- chart_type: line
- dimensions:
- - name: faults
- - name: perf.l1d_cache
- description: L1D cache operations
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: read_access
- - name: read_misses
- - name: write_access
- - name: write_misses
- - name: perf.l1d_cache_prefetch
- description: L1D prefetch cache operations
- unit: "prefetches/s"
- chart_type: line
- dimensions:
- - name: prefetches
- - name: perf.l1i_cache
- description: L1I cache operations
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: read_access
- - name: read_misses
- - name: perf.ll_cache
- description: LL cache operations
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: read_access
- - name: read_misses
- - name: write_access
- - name: write_misses
- - name: perf.dtlb_cache
- description: DTLB cache operations
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: read_access
- - name: read_misses
- - name: write_access
- - name: write_misses
- - name: perf.itlb_cache
- description: ITLB cache operations
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: read_access
- - name: read_misses
- - name: perf.pbu_cache
- description: PBU cache operations
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: read_access
diff --git a/collectors/plugins.d/Makefile.am b/collectors/plugins.d/Makefile.am
deleted file mode 100644
index 67fed309d..000000000
--- a/collectors/plugins.d/Makefile.am
+++ /dev/null
@@ -1,12 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-SUBDIRS = \
- $(NULL)
-
-dist_noinst_DATA = \
- gperf-config.txt \
- README.md \
- $(NULL)
diff --git a/collectors/plugins.d/README.md b/collectors/plugins.d/README.md
deleted file mode 100644
index 0752d389b..000000000
--- a/collectors/plugins.d/README.md
+++ /dev/null
@@ -1,680 +0,0 @@
-<!--
-title: "External plugins"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/plugins.d/README.md"
-sidebar_label: "External plugins"
-learn_status: "Published"
-learn_topic_type: "References"
-learn_rel_path: "Developers/External plugins"
--->
-
-# External plugins
-
-`plugins.d` is the Netdata internal plugin that collects metrics
-from external processes, thus allowing Netdata to use **external plugins**.
-
-## Provided External Plugins
-
-| plugin | language | O/S | description |
-|:------------------------------------------------------------------------------------------------------:|:--------:|:--------------:|:----------------------------------------------------------------------------------------------------------------------------------------|
-| [apps.plugin](https://github.com/netdata/netdata/blob/master/collectors/apps.plugin/README.md) | `C` | linux, freebsd | monitors the whole process tree on Linux and FreeBSD and breaks down system resource usage by **process**, **user** and **user group**. |
-| [charts.d.plugin](https://github.com/netdata/netdata/blob/master/collectors/charts.d.plugin/README.md) | `BASH` | all | a **plugin orchestrator** for data collection modules written in `BASH` v4+. |
-| [cups.plugin](https://github.com/netdata/netdata/blob/master/collectors/cups.plugin/README.md) | `C` | all | monitors **CUPS** |
-| [ebpf.plugin](https://github.com/netdata/netdata/blob/master/collectors/ebpf.plugin/README.md) | `C` | linux | monitors different metrics on environments using kernel internal functions. |
-| [go.d.plugin](https://github.com/netdata/go.d.plugin/blob/master/README.md) | `GO` | all | collects metrics from the system, applications, or third-party APIs. |
-| [ioping.plugin](https://github.com/netdata/netdata/blob/master/collectors/ioping.plugin/README.md) | `C` | all | measures disk latency. |
-| [freeipmi.plugin](https://github.com/netdata/netdata/blob/master/collectors/freeipmi.plugin/README.md) | `C` | linux | collects metrics from enterprise hardware sensors, on Linux servers. |
-| [nfacct.plugin](https://github.com/netdata/netdata/blob/master/collectors/nfacct.plugin/README.md) | `C` | linux | collects netfilter firewall, connection tracker and accounting metrics using `libmnl` and `libnetfilter_acct`. |
-| [xenstat.plugin](https://github.com/netdata/netdata/blob/master/collectors/xenstat.plugin/README.md) | `C` | linux | collects XenServer and XCP-ng metrics using `lxenstat`. |
-| [perf.plugin](https://github.com/netdata/netdata/blob/master/collectors/perf.plugin/README.md) | `C` | linux | collects CPU performance metrics using performance monitoring units (PMU). |
-| [python.d.plugin](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/README.md) | `python` | all | a **plugin orchestrator** for data collection modules written in `python` v2 or v3 (both are supported). |
-| [slabinfo.plugin](https://github.com/netdata/netdata/blob/master/collectors/slabinfo.plugin/README.md) | `C` | linux | collects kernel internal cache objects (SLAB) metrics. |
-
-Plugin orchestrators may also be described as **modular plugins**. They are modular since they accept custom made modules to be included. Writing modules for these plugins is easier than accessing the native Netdata API directly. You will find modules already available for each orchestrator under the directory of the particular modular plugin (e.g. under python.d.plugin for the python orchestrator).
-Each of these modular plugins has each own methods for defining modules. Please check the examples and their documentation.
-
-## Motivation
-
-This plugin allows Netdata to use **external plugins** for data collection:
-
-1. external data collection plugins may be written in any computer language.
-
-2. external data collection plugins may use O/S capabilities or `setuid` to
- run with escalated privileges (compared to the `netdata` daemon).
- The communication between the external plugin and Netdata is unidirectional
- (from the plugin to Netdata), so that Netdata cannot manipulate an external
- plugin running with escalated privileges.
-
-## Operation
-
-Each of the external plugins is expected to run forever.
-Netdata will start it when it starts and stop it when it exits.
-
-If the external plugin exits or crashes, Netdata will log an error.
-If the external plugin exits or crashes without pushing metrics to Netdata, Netdata will not start it again.
-
-- Plugins that exit with any value other than zero, will be disabled. Plugins that exit with zero, will be restarted after some time.
-- Plugins may also be disabled by Netdata if they output things that Netdata does not understand.
-
-The `stdout` of external plugins is connected to Netdata to receive metrics,
-with the API defined below.
-
-The `stderr` of external plugins is connected to Netdata's `error.log`.
-
-Plugins can create any number of charts with any number of dimensions each. Each chart can have its own characteristics independently of the others generated by the same plugin. For example, one chart may have an update frequency of 1 second, another may have 5 seconds and a third may have 10 seconds.
-
-## Configuration
-
-Netdata will supply the environment variables `NETDATA_USER_CONFIG_DIR` (for user supplied) and `NETDATA_STOCK_CONFIG_DIR` (for Netdata supplied) configuration files to identify the directory where configuration files are stored. It is up to the plugin to read the configuration it needs.
-
-The `netdata.conf` section `[plugins]` section contains a list of all the plugins found at the system where Netdata runs, with a boolean setting to enable them or not.
-
-Example:
-
-```
-[plugins]
- # enable running new plugins = yes
- # check for new plugins every = 60
-
- # charts.d = yes
- # ioping = yes
- # python.d = yes
-```
-
-The setting `enable running new plugins` sets the default behavior for all external plugins. It can be
-overridden for distinct plugins by modifying the appropriate plugin value configuration to either `yes` or `no`.
-
-The setting `check for new plugins every` sets the interval between scans of the directory
-`/usr/libexec/netdata/plugins.d`. New plugins can be added any time, and Netdata will detect them in a timely manner.
-
-For each of the external plugins enabled, another `netdata.conf` section
-is created, in the form of `[plugin:NAME]`, where `NAME` is the name of the external plugin.
-This section allows controlling the update frequency of the plugin and provide
-additional command line arguments to it.
-
-For example, for `apps.plugin` the following section is available:
-
-```
-[plugin:apps]
- # update every = 1
- # command options =
-```
-
-- `update every` controls the granularity of the external plugin.
-- `command options` allows giving additional command line options to the plugin.
-
-Netdata will provide to the external plugins the environment variable `NETDATA_UPDATE_EVERY`, in seconds (the default is 1). This is the **minimum update frequency** for all charts. A plugin that is updating values more frequently than this, is just wasting resources.
-
-Netdata will call the plugin with just one command line parameter: the number of seconds the user requested this plugin to update its data (by default is also 1).
-
-Other than the above, the plugin configuration is up to the plugin.
-
-Keep in mind, that the user may use Netdata configuration to overwrite chart and dimension parameters. This is transparent to the plugin.
-
-### Autoconfiguration
-
-Plugins should attempt to autoconfigure themselves when possible.
-
-For example, if your plugin wants to monitor `squid`, you can search for it on port `3128` or `8080`. If any succeeds, you can proceed. If it fails you can output an error (on stderr) saying that you cannot find `squid` running and giving instructions about the plugin configuration. Then you can stop (exit with non-zero value), so that Netdata will not attempt to start the plugin again.
-
-## External Plugins API
-
-Any program that can print a few values to its standard output can become a Netdata external plugin.
-
-Netdata parses lines starting with:
-
-- `CHART` - create or update a chart
-- `DIMENSION` - add or update a dimension to the chart just created
-- `VARIABLE` - define a variable (to be used in health calculations)
-- `CLABEL` - add a label to a chart
-- `CLABEL_COMMIT` - commit added labels to the chart
-- `FUNCTION` - define a function that can be called later to execute it
-- `BEGIN` - initialize data collection for a chart
-- `SET` - set the value of a dimension for the initialized chart
-- `END` - complete data collection for the initialized chart
-- `FLUSH` - ignore the last collected values
-- `DISABLE` - disable this plugin
-
-a single program can produce any number of charts with any number of dimensions each.
-
-Charts can be added any time (not just the beginning).
-
-### Command line parameters
-
-The plugin **MUST** accept just **one** parameter: **the number of seconds it is
-expected to update the values for its charts**. The value passed by Netdata
-to the plugin is controlled via its configuration file (so there is no need
-for the plugin to handle this configuration option).
-
-The external plugin can overwrite the update frequency. For example, the server may
-request per second updates, but the plugin may ignore it and update its charts
-every 5 seconds.
-
-### Environment variables
-
-There are a few environment variables that are set by `netdata` and are
-available for the plugin to use.
-
-| variable | description |
-|:---------------------------:|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
-| `NETDATA_USER_CONFIG_DIR` | The directory where all Netdata-related user configuration should be stored. If the plugin requires custom user configuration, this is the place the user has saved it (normally under `/etc/netdata`). |
-| `NETDATA_STOCK_CONFIG_DIR` | The directory where all Netdata -related stock configuration should be stored. If the plugin is shipped with configuration files, this is the place they can be found (normally under `/usr/lib/netdata/conf.d`). |
-| `NETDATA_PLUGINS_DIR` | The directory where all Netdata plugins are stored. |
-| `NETDATA_USER_PLUGINS_DIRS` | The list of directories where custom plugins are stored. |
-| `NETDATA_WEB_DIR` | The directory where the web files of Netdata are saved. |
-| `NETDATA_CACHE_DIR` | The directory where the cache files of Netdata are stored. Use this directory if the plugin requires a place to store data. A new directory should be created for the plugin for this purpose, inside this directory. |
-| `NETDATA_LOG_DIR` | The directory where the log files are stored. By default the `stderr` output of the plugin will be saved in the `error.log` file of Netdata. |
-| `NETDATA_HOST_PREFIX` | This is used in environments where system directories like `/sys` and `/proc` have to be accessed at a different path. |
-| `NETDATA_DEBUG_FLAGS` | This is a number (probably in hex starting with `0x`), that enables certain Netdata debugging features. Check **\[[Tracing Options]]** for more information. |
-| `NETDATA_UPDATE_EVERY` | The minimum number of seconds between chart refreshes. This is like the **internal clock** of Netdata (it is user configurable, defaulting to `1`). There is no meaning for a plugin to update its values more frequently than this number of seconds. |
-
-### The output of the plugin
-
-The plugin should output instructions for Netdata to its output (`stdout`). Since this uses pipes, please make sure you flush stdout after every iteration.
-
-#### DISABLE
-
-`DISABLE` will disable this plugin. This will prevent Netdata from restarting the plugin. You can also exit with the value `1` to have the same effect.
-
-#### HOST_DEFINE
-
-`HOST_DEFINE` defines a new (or updates an existing) virtual host.
-
-The template is:
-
-> HOST_DEFINE machine_guid hostname
-
-where:
-
-- `machine_guid`
-
- uniquely identifies the host, this is what will be needed to add charts to the host.
-
-- `hostname`
-
- is the hostname of the virtual host
-
-#### HOST_LABEL
-
-`HOST_LABEL` adds a key-value pair to the virtual host labels. It has to be given between `HOST_DEFINE` and `HOST_DEFINE_END`.
-
-The template is:
-
-> HOST_LABEL key value
-
-where:
-
-- `key`
-
- uniquely identifies the key of the label
-
-- `value`
-
- is the value associated with this key
-
-There are a few special keys that are used to define the system information of the monitored system:
-
-- `_cloud_provider_type`
-- `_cloud_instance_type`
-- `_cloud_instance_region`
-- `_os_name`
-- `_os_version`
-- `_kernel_version`
-- `_system_cores`
-- `_system_cpu_freq`
-- `_system_ram_total`
-- `_system_disk_space`
-- `_architecture`
-- `_virtualization`
-- `_container`
-- `_container_detection`
-- `_virt_detection`
-- `_is_k8s_node`
-- `_install_type`
-- `_prebuilt_arch`
-- `_prebuilt_dist`
-
-#### HOST_DEFINE_END
-
-`HOST_DEFINE_END` commits the host information, creating a new host entity, or updating an existing one with the same `machine_guid`.
-
-#### HOST
-
-`HOST` switches data collection between hosts.
-
-The template is:
-
-> HOST machine_guid
-
-where:
-
-- `machine_guid`
-
- is the UUID of the host to switch to. After this command, every other command following it is assumed to be associated with this host.
- Setting machine_guid to `localhost` switches data collection to the local host.
-
-#### CHART
-
-`CHART` defines a new chart.
-
-the template is:
-
-> CHART type.id name title units \[family \[context \[charttype \[priority \[update_every \[options \[plugin [module]]]]]]]]
-
- where:
-
-- `type.id`
-
- uniquely identifies the chart,
- this is what will be needed to add values to the chart
-
- the `type` part controls the menu the charts will appear in
-
-- `name`
-
- is the name that will be presented to the user instead of `id` in `type.id`. This means that only the `id` part of
- `type.id` is changed. When a name has been given, the chart is indexed (and can be referred) as both `type.id` and
- `type.name`. You can set name to `''`, or `null`, or `(null)` to disable it. If a chart with the same name already
- exists, a serial number is automatically attached to the name to avoid naming collisions.
-
-- `title`
-
- the text above the chart
-
-- `units`
-
- the label of the vertical axis of the chart,
- all dimensions added to a chart should have the same units
- of measurement
-
-- `family`
-
- is used to group charts together
- (for example all eth0 charts should say: eth0),
- if empty or missing, the `id` part of `type.id` will be used
-
- this controls the sub-menu on the dashboard
-
-- `context`
-
- the context is giving the template of the chart. For example, if multiple charts present the same information for a different family, they should have the same `context`
-
- this is used for looking up rendering information for the chart (colors, sizes, informational texts) and also apply alerts to it
-
-- `charttype`
-
- one of `line`, `area` or `stacked`,
- if empty or missing, the `line` will be used
-
-- `priority`
-
- is the relative priority of the charts as rendered on the web page,
- lower numbers make the charts appear before the ones with higher numbers,
- if empty or missing, `1000` will be used
-
-- `update_every`
-
- overwrite the update frequency set by the server,
- if empty or missing, the user configured value will be used
-
-- `options`
-
- a space separated list of options, enclosed in quotes. 4 options are currently supported: `obsolete` to mark a chart as obsolete (Netdata will hide it and delete it after some time), `detail` to mark a chart as insignificant (this may be used by dashboards to make the charts smaller, or somehow visualize properly a less important chart), `store_first` to make Netdata store the first collected value, assuming there was an invisible previous value set to zero (this is used by statsd charts - if the first data collected value of incremental dimensions is not zero based, unrealistic spikes will appear with this option set) and `hidden` to perform all operations on a chart, but do not offer it on dashboards (the chart will be send to external databases). `CHART` options have been added in Netdata v1.7 and the `hidden` option was added in 1.10.
-
-- `plugin` and `module`
-
- both are just names that are used to let the user identify the plugin and the module that generated the chart. If `plugin` is unset or empty, Netdata will automatically set the filename of the plugin that generated the chart. `module` has not default.
-
-#### DIMENSION
-
-`DIMENSION` defines a new dimension for the chart
-
-the template is:
-
-> DIMENSION id \[name \[algorithm \[multiplier \[divisor [options]]]]]
-
- where:
-
-- `id`
-
- the `id` of this dimension (it is a text value, not numeric),
- this will be needed later to add values to the dimension
-
- We suggest to avoid using `.` in dimension ids. External databases expect metrics to be `.` separated and people will get confused if a dimension id contains a dot.
-
-- `name`
-
- the name of the dimension as it will appear at the legend of the chart,
- if empty or missing the `id` will be used
-
-- `algorithm`
-
- one of:
-
- - `absolute`
-
- the value is to drawn as-is (interpolated to second boundary),
- if `algorithm` is empty, invalid or missing, `absolute` is used
-
- - `incremental`
-
- the value increases over time,
- the difference from the last value is presented in the chart,
- the server interpolates the value and calculates a per second figure
-
- - `percentage-of-absolute-row`
-
- the % of this value compared to the total of all dimensions
-
- - `percentage-of-incremental-row`
-
- the % of this value compared to the incremental total of
- all dimensions
-
-- `multiplier`
-
- an integer value to multiply the collected value,
- if empty or missing, `1` is used
-
-- `divisor`
-
- an integer value to divide the collected value,
- if empty or missing, `1` is used
-
-- `options`
-
- a space separated list of options, enclosed in quotes. Options supported: `obsolete` to mark a dimension as obsolete (Netdata will delete it after some time) and `hidden` to make this dimension hidden, it will take part in the calculations but will not be presented in the chart.
-
-#### VARIABLE
-
-> VARIABLE [SCOPE] name = value
-
-`VARIABLE` defines a variable that can be used in alerts. This is to used for setting constants (like the max connections a server may accept).
-
-Variables support 2 scopes:
-
-- `GLOBAL` or `HOST` to define the variable at the host level.
-- `LOCAL` or `CHART` to define the variable at the chart level. Use chart-local variables when the same variable may exist for different charts (i.e. Netdata monitors 2 mysql servers, and you need to set the `max_connections` each server accepts). Using chart-local variables is the ideal to build alert templates.
-
-The position of the `VARIABLE` line, sets its default scope (in case you do not specify a scope). So, defining a `VARIABLE` before any `CHART`, or between `END` and `BEGIN` (outside any chart), sets `GLOBAL` scope, while defining a `VARIABLE` just after a `CHART` or a `DIMENSION`, or within the `BEGIN` - `END` block of a chart, sets `LOCAL` scope.
-
-These variables can be set and updated at any point.
-
-Variable names should use alphanumeric characters, the `.` and the `_`.
-
-The `value` is floating point (Netdata used `long double`).
-
-Variables are transferred to upstream Netdata servers (streaming and database replication).
-
-#### CLABEL
-
-> CLABEL name value source
-
-`CLABEL` defines a label used to organize and identify a chart.
-
-Name and value accept characters according to the following table:
-
-| Character | Symbol | Label Name | Label Value |
-|---------------------|:------:|:----------:|:-----------:|
-| UTF-8 character | UTF-8 | _ | keep |
-| Lower case letter | [a-z] | keep | keep |
-| Upper case letter | [A-Z] | keep | [a-z] |
-| Digit | [0-9] | keep | keep |
-| Underscore | _ | keep | keep |
-| Minus | - | keep | keep |
-| Plus | + | _ | keep |
-| Colon | : | _ | keep |
-| Semicolon | ; | _ | : |
-| Equal | = | _ | : |
-| Period | . | keep | keep |
-| Comma | , | . | . |
-| Slash | / | keep | keep |
-| Backslash | \ | / | / |
-| At | @ | _ | keep |
-| Space | ' ' | _ | keep |
-| Opening parenthesis | ( | _ | keep |
-| Closing parenthesis | ) | _ | keep |
-| Anything else | | _ | _ |
-
-The `source` is an integer field that can have the following values:
-- `1`: The value was set automatically.
-- `2`: The value was set manually.
-- `4`: This is a K8 label.
-- `8`: This is a label defined using `netdata` agent cloud link.
-
-#### CLABEL_COMMIT
-
-`CLABEL_COMMIT` indicates that all labels were defined and the chart can be updated.
-
-#### FUNCTION
-
-> FUNCTION [GLOBAL] "name and parameters of the function" timeout "help string for users"
-
-A function can be used by users to ask for more information from the collector. Netdata maintains a registry of functions in 2 levels:
-
-- per node
-- per chart
-
-Both node and chart functions are exactly the same, but chart functions allow Netdata to relate functions with charts and therefore present a context sensitive menu of functions related to the chart the user is using.
-
-A function is identified by a string. The allowed characters in the function definition are:
-
-| Character | Symbol | In Functions |
-|-------------------|:------:|:------------:|
-| UTF-8 character | UTF-8 | keep |
-| Lower case letter | [a-z] | keep |
-| Upper case letter | [A-Z] | keep |
-| Digit | [0-9] | keep |
-| Underscore | _ | keep |
-| Comma | , | keep |
-| Minus | - | keep |
-| Period | . | keep |
-| Colon | : | keep |
-| Slash | / | keep |
-| Space | ' ' | keep |
-| Semicolon | ; | : |
-| Equal | = | : |
-| Backslash | \ | / |
-| Anything else | | _ |
-
-Uses can get a list of all the registered functions using the `/api/v1/functions` end point of Netdata.
-
-Users can call functions using the `/api/v1/function` end point of Netdata.
-Once a function is called, the plugin will receive at its standard input a command that looks like this:
-
-> FUNCTION transaction_id timeout "name and parameters of the function"
-
-The plugin is expected to parse and validate `name and parameters of the function`. Netdata allows users to edit this string, append more parameters or even change the ones the plugin originally exposed. To minimize the security risk, Netdata guarantees that only the characters shown above are accepted in function definitions, but still the plugin should carefully inspect the `name and parameters of the function` to ensure that it is valid and not harmful.
-
-If the plugin rejects the request, it should respond with this:
-
-```
-FUNCTION_RESULT_BEGIN transaction_id 400 application/json
-{
- "status": 400,
- "error_message": "description of the rejection reasons"
-}
-FUNCTION_RESULT_END
-```
-
-If the plugin prepares a response, it should send (via its standard output, together with the collected data, but not interleaved with them):
-
-> FUNCTION_RESULT_BEGIN transaction_id http_error_code content_type expiration
-
-Where:
-
- - `transaction_id` is the transaction id that Netdata sent for this function execution
- - `http_error` is the http error code Netdata should respond with, 200 is the "ok" response
- - `content_type` is the content type of the response
- - `expiration` is the absolute timestamp (number, unix epoch) this response expires
-
-Immediately after this, all text is assumed to be the response content.
-The content is text and line oriented. The maximum line length accepted is 15kb. Longer lines will be truncated.
-The type of the context itself depends on the plugin and the UI.
-
-To terminate the message, Netdata seeks a line with just this:
-
-> FUNCTION_RESULT_END
-
-This defines the end of the message. `FUNCTION_RESULT_END` should appear in a line alone, without any other text, so it is wise to add `\n` before and after it.
-
-After this line, Netdata resumes processing collected metrics from the plugin.
-
-## Data collection
-
-data collection is defined as a series of `BEGIN` -> `SET` -> `END` lines
-
-> BEGIN type.id [microseconds]
-
-- `type.id`
-
- is the unique identification of the chart (as given in `CHART`)
-
-- `microseconds`
-
- is the number of microseconds since the last update of the chart. It is optional.
-
- Under heavy system load, the system may have some latency transferring
- data from the plugins to Netdata via the pipe. This number improves
- accuracy significantly, since the plugin is able to calculate the
- duration between its iterations better than Netdata.
-
- The first time the plugin is started, no microseconds should be given
- to Netdata.
-
-> SET id = value
-
-- `id`
-
- is the unique identification of the dimension (of the chart just began)
-
-- `value`
-
- is the collected value, only integer values are collected. If you want to push fractional values, multiply this value by 100 or 1000 and set the `DIMENSION` divider to 1000.
-
-> END
-
- END does not take any parameters, it commits the collected values for all dimensions to the chart. If a dimensions was not `SET`, its value will be empty for this commit.
-
-More `SET` lines may appear to update all the dimensions of the chart.
-All of them in one `BEGIN` -> `END` block.
-
-All `SET` lines within a single `BEGIN` -> `END` block have to refer to the
-same chart.
-
-If more charts need to be updated, each chart should have its own
-`BEGIN` -> `SET` -> `END` block.
-
-If, for any reason, a plugin has issued a `BEGIN` but wants to cancel it,
-it can issue a `FLUSH`. The `FLUSH` command will instruct Netdata to ignore
-all the values collected since the last `BEGIN` command.
-
-If a plugin does not behave properly (outputs invalid lines, or does not
-follow these guidelines), will be disabled by Netdata.
-
-### collected values
-
-Netdata will collect any **signed** value in the 64bit range:
-`-9.223.372.036.854.775.808` to `+9.223.372.036.854.775.807`
-
-If a value is not collected, leave it empty, like this:
-
-`SET id =`
-
-or do not output the line at all.
-
-## Modular Plugins
-
-1. **python**, use `python.d.plugin`, there are many examples in the [python.d
- directory](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/README.md)
-
- python is ideal for Netdata plugins. It is a simple, yet powerful way to collect data, it has a very small memory footprint, although it is not the most CPU efficient way to do it.
-
-2. **BASH**, use `charts.d.plugin`, there are many examples in the [charts.d
- directory](https://github.com/netdata/netdata/blob/master/collectors/charts.d.plugin/README.md)
-
- BASH is the simplest scripting language for collecting values. It is the less efficient though in terms of CPU resources. You can use it to collect data quickly, but extensive use of it might use a lot of system resources.
-
-3. **C**
-
- Of course, C is the most efficient way of collecting data. This is why Netdata itself is written in C.
-
-## Writing Plugins Properly
-
-There are a few rules for writing plugins properly:
-
-1. Respect system resources
-
- Pay special attention to efficiency:
-
- - Initialize everything once, at the beginning. Initialization is not an expensive operation. Your plugin will most probably be started once and run forever. So, do whatever heavy operation is needed at the beginning, just once.
- - Do the absolutely minimum while iterating to collect values repeatedly.
- - If you need to connect to another server to collect values, avoid re-connects if possible. Connect just once, with keep-alive (for HTTP) enabled and collect values using the same connection.
- - Avoid any CPU or memory heavy operation while collecting data. If you control memory allocation, avoid any memory allocation while iterating to collect values.
- - Avoid running external commands when possible. If you are writing shell scripts avoid especially pipes (each pipe is another fork, a very expensive operation).
-
-2. The best way to iterate at a constant pace is this pseudo code:
-
-```js
- var update_every = argv[1] * 1000; /* seconds * 1000 = milliseconds */
-
- readConfiguration();
-
- if(!verifyWeCanCollectValues()) {
- print("DISABLE");
- exit(1);
- }
-
- createCharts(); /* print CHART and DIMENSION statements */
-
- var loops = 0;
- var last_run = 0;
- var next_run = 0;
- var dt_since_last_run = 0;
- var now = 0;
-
- while(true) {
- /* find the current time in milliseconds */
- now = currentTimeStampInMilliseconds();
-
- /*
- * find the time of the next loop
- * this makes sure we are always aligned
- * with the Netdata daemon
- */
- next_run = now - (now % update_every) + update_every;
-
- /*
- * wait until it is time
- * it is important to do it in a loop
- * since many wait functions can be interrupted
- */
- while( now < next_run ) {
- sleepMilliseconds(next_run - now);
- now = currentTimeStampInMilliseconds();
- }
-
- /* calculate the time passed since the last run */
- if ( loops > 0 )
- dt_since_last_run = (now - last_run) * 1000; /* in microseconds */
-
- /* prepare for the next loop */
- last_run = now;
- loops++;
-
- /* do your magic here to collect values */
- collectValues();
-
- /* send the collected data to Netdata */
- printValues(dt_since_last_run); /* print BEGIN, SET, END statements */
- }
-```
-
- Using the above procedure, your plugin will be synchronized to start data collection on steps of `update_every`. There will be no need to keep track of latencies in data collection.
-
- Netdata interpolates values to second boundaries, so even if your plugin is not perfectly aligned it does not matter. Netdata will find out. When your plugin works in increments of `update_every`, there will be no gaps in the charts due to the possible cumulative micro-delays in data collection. Gaps will only appear if the data collection is really delayed.
-
-3. If you are not sure of memory leaks, exit every one hour. Netdata will re-start your process.
-
-4. If possible, try to autodetect if your plugin should be enabled, without any configuration.
-
-
diff --git a/collectors/plugins.d/gperf-config.txt b/collectors/plugins.d/gperf-config.txt
deleted file mode 100644
index bad51367c..000000000
--- a/collectors/plugins.d/gperf-config.txt
+++ /dev/null
@@ -1,58 +0,0 @@
-%struct-type
-%omit-struct-type
-%define hash-function-name gperf_keyword_hash_function
-%define lookup-function-name gperf_lookup_keyword
-%define word-array-name gperf_keywords
-%define constants-prefix GPERF_PARSER_
-%define slot-name keyword
-%global-table
-%null-strings
-PARSER_KEYWORD;
-%%
-#
-# Plugins Only Keywords
-#
-FLUSH, 97, PARSER_INIT_PLUGINSD, WORKER_PARSER_FIRST_JOB + 1
-DISABLE, 98, PARSER_INIT_PLUGINSD, WORKER_PARSER_FIRST_JOB + 2
-EXIT, 99, PARSER_INIT_PLUGINSD, WORKER_PARSER_FIRST_JOB + 3
-HOST, 71, PARSER_INIT_PLUGINSD|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 4
-HOST_DEFINE, 72, PARSER_INIT_PLUGINSD|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 5
-HOST_DEFINE_END, 73, PARSER_INIT_PLUGINSD|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 6
-HOST_LABEL, 74, PARSER_INIT_PLUGINSD|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 7
-#
-# Common keywords
-#
-BEGIN, 12, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 8
-CHART, 32, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 9
-CLABEL, 34, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 10
-CLABEL_COMMIT, 35, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 11
-DIMENSION, 31, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 12
-END, 13, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 13
-FUNCTION, 41, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 14
-FUNCTION_RESULT_BEGIN, 42, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 15
-LABEL, 51, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 16
-OVERWRITE, 52, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 17
-SET, 11, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 18
-VARIABLE, 53, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 19
-DYNCFG_ENABLE, 101, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 20
-DYNCFG_REGISTER_MODULE, 102, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 21
-DYNCFG_REGISTER_JOB, 103, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 22
-DYNCFG_RESET, 104, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 23
-REPORT_JOB_STATUS, 110, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 24
-DELETE_JOB, 111, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 25
-#
-# Streaming only keywords
-#
-CLAIMED_ID, 61, PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 26
-BEGIN2, 2, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 27
-SET2, 1, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 28
-END2, 3, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 29
-#
-# Streaming Replication keywords
-#
-CHART_DEFINITION_END, 33, PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 30
-RBEGIN, 22, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 31
-RDSTATE, 23, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 32
-REND, 25, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 33
-RSET, 21, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 34
-RSSTATE, 24, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 35
diff --git a/collectors/plugins.d/gperf-hashtable.h b/collectors/plugins.d/gperf-hashtable.h
deleted file mode 100644
index b327d8d6d..000000000
--- a/collectors/plugins.d/gperf-hashtable.h
+++ /dev/null
@@ -1,177 +0,0 @@
-/* ANSI-C code produced by gperf version 3.1 */
-/* Command-line: gperf --multiple-iterations=1000 --output-file=gperf-hashtable.h gperf-config.txt */
-/* Computed positions: -k'1-2' */
-
-#if !((' ' == 32) && ('!' == 33) && ('"' == 34) && ('#' == 35) \
- && ('%' == 37) && ('&' == 38) && ('\'' == 39) && ('(' == 40) \
- && (')' == 41) && ('*' == 42) && ('+' == 43) && (',' == 44) \
- && ('-' == 45) && ('.' == 46) && ('/' == 47) && ('0' == 48) \
- && ('1' == 49) && ('2' == 50) && ('3' == 51) && ('4' == 52) \
- && ('5' == 53) && ('6' == 54) && ('7' == 55) && ('8' == 56) \
- && ('9' == 57) && (':' == 58) && (';' == 59) && ('<' == 60) \
- && ('=' == 61) && ('>' == 62) && ('?' == 63) && ('A' == 65) \
- && ('B' == 66) && ('C' == 67) && ('D' == 68) && ('E' == 69) \
- && ('F' == 70) && ('G' == 71) && ('H' == 72) && ('I' == 73) \
- && ('J' == 74) && ('K' == 75) && ('L' == 76) && ('M' == 77) \
- && ('N' == 78) && ('O' == 79) && ('P' == 80) && ('Q' == 81) \
- && ('R' == 82) && ('S' == 83) && ('T' == 84) && ('U' == 85) \
- && ('V' == 86) && ('W' == 87) && ('X' == 88) && ('Y' == 89) \
- && ('Z' == 90) && ('[' == 91) && ('\\' == 92) && (']' == 93) \
- && ('^' == 94) && ('_' == 95) && ('a' == 97) && ('b' == 98) \
- && ('c' == 99) && ('d' == 100) && ('e' == 101) && ('f' == 102) \
- && ('g' == 103) && ('h' == 104) && ('i' == 105) && ('j' == 106) \
- && ('k' == 107) && ('l' == 108) && ('m' == 109) && ('n' == 110) \
- && ('o' == 111) && ('p' == 112) && ('q' == 113) && ('r' == 114) \
- && ('s' == 115) && ('t' == 116) && ('u' == 117) && ('v' == 118) \
- && ('w' == 119) && ('x' == 120) && ('y' == 121) && ('z' == 122) \
- && ('{' == 123) && ('|' == 124) && ('}' == 125) && ('~' == 126))
-/* The character set is not based on ISO-646. */
-#error "gperf generated tables don't work with this execution character set. Please report a bug to <bug-gperf@gnu.org>."
-#endif
-
-
-#define GPERF_PARSER_TOTAL_KEYWORDS 35
-#define GPERF_PARSER_MIN_WORD_LENGTH 3
-#define GPERF_PARSER_MAX_WORD_LENGTH 22
-#define GPERF_PARSER_MIN_HASH_VALUE 3
-#define GPERF_PARSER_MAX_HASH_VALUE 47
-/* maximum key range = 45, duplicates = 0 */
-
-#ifdef __GNUC__
-__inline
-#else
-#ifdef __cplusplus
-inline
-#endif
-#endif
-static unsigned int
-gperf_keyword_hash_function (register const char *str, register size_t len)
-{
- static unsigned char asso_values[] =
- {
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 11, 18, 0, 0, 0,
- 6, 48, 9, 0, 48, 48, 20, 48, 0, 8,
- 48, 48, 1, 12, 48, 20, 18, 48, 2, 0,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48, 48, 48, 48, 48,
- 48, 48, 48, 48, 48, 48
- };
- return len + asso_values[(unsigned char)str[1]] + asso_values[(unsigned char)str[0]];
-}
-
-static PARSER_KEYWORD gperf_keywords[] =
- {
- {(char*)0}, {(char*)0}, {(char*)0},
-#line 30 "gperf-config.txt"
- {"END", 13, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 13},
-#line 49 "gperf-config.txt"
- {"END2", 3, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 29},
-#line 56 "gperf-config.txt"
- {"REND", 25, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 33},
-#line 17 "gperf-config.txt"
- {"EXIT", 99, PARSER_INIT_PLUGINSD, WORKER_PARSER_FIRST_JOB + 3},
-#line 16 "gperf-config.txt"
- {"DISABLE", 98, PARSER_INIT_PLUGINSD, WORKER_PARSER_FIRST_JOB + 2},
-#line 55 "gperf-config.txt"
- {"RDSTATE", 23, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 32},
-#line 29 "gperf-config.txt"
- {"DIMENSION", 31, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 12},
-#line 42 "gperf-config.txt"
- {"DELETE_JOB", 111, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 25},
- {(char*)0},
-#line 40 "gperf-config.txt"
- {"DYNCFG_RESET", 104, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 23},
-#line 37 "gperf-config.txt"
- {"DYNCFG_ENABLE", 101, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 20},
-#line 26 "gperf-config.txt"
- {"CHART", 32, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 9},
-#line 35 "gperf-config.txt"
- {"SET", 11, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 18},
-#line 48 "gperf-config.txt"
- {"SET2", 1, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 28},
-#line 57 "gperf-config.txt"
- {"RSET", 21, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 34},
-#line 41 "gperf-config.txt"
- {"REPORT_JOB_STATUS", 110, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 24},
-#line 39 "gperf-config.txt"
- {"DYNCFG_REGISTER_JOB", 103, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 22},
-#line 58 "gperf-config.txt"
- {"RSSTATE", 24, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 35},
-#line 18 "gperf-config.txt"
- {"HOST", 71, PARSER_INIT_PLUGINSD|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 4},
-#line 38 "gperf-config.txt"
- {"DYNCFG_REGISTER_MODULE", 102, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 21},
-#line 25 "gperf-config.txt"
- {"BEGIN", 12, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 8},
-#line 47 "gperf-config.txt"
- {"BEGIN2", 2, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 27},
-#line 54 "gperf-config.txt"
- {"RBEGIN", 22, PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 31},
-#line 27 "gperf-config.txt"
- {"CLABEL", 34, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 10},
-#line 21 "gperf-config.txt"
- {"HOST_LABEL", 74, PARSER_INIT_PLUGINSD|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 7},
-#line 19 "gperf-config.txt"
- {"HOST_DEFINE", 72, PARSER_INIT_PLUGINSD|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 5},
-#line 53 "gperf-config.txt"
- {"CHART_DEFINITION_END", 33, PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 30},
-#line 46 "gperf-config.txt"
- {"CLAIMED_ID", 61, PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 26},
-#line 15 "gperf-config.txt"
- {"FLUSH", 97, PARSER_INIT_PLUGINSD, WORKER_PARSER_FIRST_JOB + 1},
-#line 20 "gperf-config.txt"
- {"HOST_DEFINE_END", 73, PARSER_INIT_PLUGINSD|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 6},
-#line 28 "gperf-config.txt"
- {"CLABEL_COMMIT", 35, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 11},
-#line 31 "gperf-config.txt"
- {"FUNCTION", 41, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 14},
-#line 34 "gperf-config.txt"
- {"OVERWRITE", 52, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 17},
-#line 33 "gperf-config.txt"
- {"LABEL", 51, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 16},
-#line 36 "gperf-config.txt"
- {"VARIABLE", 53, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING|PARSER_REP_METADATA, WORKER_PARSER_FIRST_JOB + 19},
- {(char*)0}, {(char*)0}, {(char*)0}, {(char*)0},
- {(char*)0}, {(char*)0}, {(char*)0}, {(char*)0},
- {(char*)0},
-#line 32 "gperf-config.txt"
- {"FUNCTION_RESULT_BEGIN", 42, PARSER_INIT_PLUGINSD|PARSER_INIT_STREAMING, WORKER_PARSER_FIRST_JOB + 15}
- };
-
-PARSER_KEYWORD *
-gperf_lookup_keyword (register const char *str, register size_t len)
-{
- if (len <= GPERF_PARSER_MAX_WORD_LENGTH && len >= GPERF_PARSER_MIN_WORD_LENGTH)
- {
- register unsigned int key = gperf_keyword_hash_function (str, len);
-
- if (key <= GPERF_PARSER_MAX_HASH_VALUE)
- {
- register const char *s = gperf_keywords[key].keyword;
-
- if (s && *str == *s && !strcmp (str + 1, s + 1))
- return &gperf_keywords[key];
- }
- }
- return 0;
-}
diff --git a/collectors/plugins.d/local_listeners.c b/collectors/plugins.d/local_listeners.c
deleted file mode 100644
index f2c5e688b..000000000
--- a/collectors/plugins.d/local_listeners.c
+++ /dev/null
@@ -1,400 +0,0 @@
-#include "libnetdata/libnetdata.h"
-#include "libnetdata/required_dummies.h"
-
-#include <stdio.h>
-#include <stdlib.h>
-#include <stdbool.h>
-#include <dirent.h>
-#include <string.h>
-#include <unistd.h>
-#include <ctype.h>
-#include <arpa/inet.h>
-
-typedef enum {
- PROC_NET_PROTOCOL_TCP,
- PROC_NET_PROTOCOL_TCP6,
- PROC_NET_PROTOCOL_UDP,
- PROC_NET_PROTOCOL_UDP6,
-} PROC_NET_PROTOCOLS;
-
-#define MAX_ERROR_LOGS 10
-
-static size_t pid_fds_processed = 0;
-static size_t pid_fds_failed = 0;
-static size_t errors_encountered = 0;
-
-static inline const char *protocol_name(PROC_NET_PROTOCOLS protocol) {
- switch(protocol) {
- default:
- case PROC_NET_PROTOCOL_TCP:
- return "TCP";
-
- case PROC_NET_PROTOCOL_UDP:
- return "UDP";
-
- case PROC_NET_PROTOCOL_TCP6:
- return "TCP6";
-
- case PROC_NET_PROTOCOL_UDP6:
- return "UDP6";
- }
-}
-
-static inline int read_cmdline(pid_t pid, char* buffer, size_t bufferSize) {
- char path[FILENAME_MAX + 1];
- snprintfz(path, FILENAME_MAX, "%s/proc/%d/cmdline", netdata_configured_host_prefix, pid);
-
- FILE* file = fopen(path, "r");
- if (!file) {
- if(++errors_encountered < MAX_ERROR_LOGS)
- collector_error("LOCAL-LISTENERS: error opening file: %s\n", path);
-
- return -1;
- }
-
- size_t bytesRead = fread(buffer, 1, bufferSize - 1, file);
- buffer[bytesRead] = '\0'; // Ensure null-terminated
-
- // Replace null characters in cmdline with spaces
- for (size_t i = 0; i < bytesRead; i++) {
- if (buffer[i] == '\0') {
- buffer[i] = ' ';
- }
- }
-
- fclose(file);
- return 0;
-}
-
-static inline void fix_cmdline(char* str) {
- if (str == NULL)
- return;
-
- char *s = str;
-
- do {
- if(*s == '|' || iscntrl(*s))
- *s = '_';
-
- } while(*++s);
-
-
- while(s > str && *(s-1) == ' ')
- *--s = '\0';
-}
-
-// ----------------------------------------------------------------------------
-
-#define HASH_TABLE_SIZE 100000
-
-typedef struct Node {
- unsigned int inode; // key
-
- // values
- unsigned int port;
- char local_address[INET6_ADDRSTRLEN];
- PROC_NET_PROTOCOLS protocol;
- bool processed;
-
- // linking
- struct Node *prev, *next;
-} Node;
-
-typedef struct HashTable {
- Node *table[HASH_TABLE_SIZE];
-} HashTable;
-
-static HashTable *hashTable_key_inode_port_value = NULL;
-
-static inline void generate_output(const char *protocol, const char *address, unsigned int port, const char *cmdline) {
- printf("%s|%s|%u|%s\n", protocol, address, port, cmdline);
-}
-
-HashTable* createHashTable() {
- HashTable *hashTable = (HashTable*)mallocz(sizeof(HashTable));
- memset(hashTable, 0, sizeof(HashTable));
- return hashTable;
-}
-
-static inline unsigned int hashFunction(unsigned int inode) {
- return inode % HASH_TABLE_SIZE;
-}
-
-static inline void insertHashTable(HashTable *hashTable, unsigned int inode, unsigned int port, PROC_NET_PROTOCOLS protocol, char *local_address) {
- unsigned int index = hashFunction(inode);
- Node *newNode = (Node*)mallocz(sizeof(Node));
- newNode->inode = inode;
- newNode->port = port;
- newNode->protocol = protocol;
- strncpyz(newNode->local_address, local_address, INET6_ADDRSTRLEN - 1);
- DOUBLE_LINKED_LIST_APPEND_ITEM_UNSAFE(hashTable->table[index], newNode, prev, next);
-}
-
-static inline bool lookupHashTable_and_execute(HashTable *hashTable, unsigned int inode, pid_t pid) {
- unsigned int index = hashFunction(inode);
- for(Node *node = hashTable->table[index], *next = NULL ; node ; node = next) {
- next = node->next;
-
- if(node->inode == inode && node->port) {
- DOUBLE_LINKED_LIST_REMOVE_ITEM_UNSAFE(hashTable->table[index], node, prev, next);
- char cmdline[8192] = "";
- read_cmdline(pid, cmdline, sizeof(cmdline));
- fix_cmdline(cmdline);
- generate_output(protocol_name(node->protocol), node->local_address, node->port, cmdline);
- freez(node);
- return true;
- }
- }
-
- return false;
-}
-
-void freeHashTable(HashTable *hashTable) {
- for (unsigned int i = 0; i < HASH_TABLE_SIZE; i++) {
- while(hashTable->table[i]) {
- Node *tmp = hashTable->table[i];
- DOUBLE_LINKED_LIST_REMOVE_ITEM_UNSAFE(hashTable->table[i], tmp, prev, next);
- generate_output(protocol_name(tmp->protocol), tmp->local_address, tmp->port, "");
- freez(tmp);
- }
- }
- freez(hashTable);
-}
-
-// ----------------------------------------------------------------------------
-
-static inline void found_this_socket_inode(pid_t pid, unsigned int inode) {
- lookupHashTable_and_execute(hashTable_key_inode_port_value, inode, pid);
-}
-
-bool find_all_sockets_in_proc(const char *proc_filename) {
- DIR *proc_dir, *fd_dir;
- struct dirent *proc_entry, *fd_entry;
- char path_buffer[FILENAME_MAX + 1];
-
- proc_dir = opendir(proc_filename);
- if (proc_dir == NULL) {
- if(++errors_encountered < MAX_ERROR_LOGS)
- collector_error("LOCAL-LISTENERS: cannot opendir() '%s'", proc_filename);
-
- pid_fds_failed++;
- return false;
- }
-
- while ((proc_entry = readdir(proc_dir)) != NULL) {
- // Check if directory entry is a PID by seeing if the name is made up of digits only
- int is_pid = 1;
- for (char *c = proc_entry->d_name; *c != '\0'; c++) {
- if (*c < '0' || *c > '9') {
- is_pid = 0;
- break;
- }
- }
-
- if (!is_pid)
- continue;
-
- // Build the path to the fd directory of the process
- snprintfz(path_buffer, FILENAME_MAX, "%s/%s/fd/", proc_filename, proc_entry->d_name);
-
- fd_dir = opendir(path_buffer);
- if (fd_dir == NULL) {
- if(++errors_encountered < MAX_ERROR_LOGS)
- collector_error("LOCAL-LISTENERS: cannot opendir() '%s'", path_buffer);
-
- pid_fds_failed++;
- continue;
- }
-
- while ((fd_entry = readdir(fd_dir)) != NULL) {
- if(!strcmp(fd_entry->d_name, ".") || !strcmp(fd_entry->d_name, ".."))
- continue;
-
- char link_path[FILENAME_MAX + 1];
- char link_target[FILENAME_MAX + 1];
- int inode;
-
- // Build the path to the file descriptor link
- snprintfz(link_path, FILENAME_MAX, "%s/%s", path_buffer, fd_entry->d_name);
-
- ssize_t len = readlink(link_path, link_target, sizeof(link_target) - 1);
- if (len == -1) {
- if(++errors_encountered < MAX_ERROR_LOGS)
- collector_error("LOCAL-LISTENERS: cannot read link '%s'", link_path);
-
- pid_fds_failed++;
- continue;
- }
- link_target[len] = '\0';
-
- pid_fds_processed++;
-
- // If the link target indicates a socket, print its inode number
- if (sscanf(link_target, "socket:[%d]", &inode) == 1)
- found_this_socket_inode((pid_t)strtoul(proc_entry->d_name, NULL, 10), inode);
- }
-
- closedir(fd_dir);
- }
-
- closedir(proc_dir);
- return true;
-}
-
-// ----------------------------------------------------------------------------
-
-static inline void add_port_and_inode(PROC_NET_PROTOCOLS protocol, unsigned int port, unsigned int inode, char *local_address) {
- insertHashTable(hashTable_key_inode_port_value, inode, port, protocol, local_address);
-}
-
-static inline void print_ipv6_address(const char *ipv6_str, char *dst) {
- unsigned k;
- char buf[9];
- struct sockaddr_in6 sa;
-
- // Initialize sockaddr_in6
- memset(&sa, 0, sizeof(struct sockaddr_in6));
- sa.sin6_family = AF_INET6;
- sa.sin6_port = htons(0); // replace 0 with your port number
-
- // Convert hex string to byte array
- for (k = 0; k < 4; ++k)
- {
- memset(buf, 0, 9);
- memcpy(buf, ipv6_str + (k * 8), 8);
- sa.sin6_addr.s6_addr32[k] = strtoul(buf, NULL, 16);
- }
-
- // Convert to human-readable format
- if (inet_ntop(AF_INET6, &(sa.sin6_addr), dst, INET6_ADDRSTRLEN) == NULL)
- *dst = '\0';
-}
-
-static inline void print_ipv4_address(uint32_t address, char *dst) {
- uint8_t octets[4];
- octets[0] = address & 0xFF;
- octets[1] = (address >> 8) & 0xFF;
- octets[2] = (address >> 16) & 0xFF;
- octets[3] = (address >> 24) & 0xFF;
- sprintf(dst, "%u.%u.%u.%u", octets[0], octets[1], octets[2], octets[3]);
-}
-
-bool read_proc_net_x(const char *filename, PROC_NET_PROTOCOLS protocol) {
- FILE *fp;
- char *line = NULL;
- size_t len = 0;
- ssize_t read;
- char address[INET6_ADDRSTRLEN];
-
- ssize_t min_line_length = (protocol == PROC_NET_PROTOCOL_TCP || protocol == PROC_NET_PROTOCOL_UDP) ? 105 : 155;
-
- fp = fopen(filename, "r");
- if (fp == NULL)
- return false;
-
- // Read line by line
- while ((read = getline(&line, &len, fp)) != -1) {
- if(read < min_line_length) continue;
-
- char local_address6[33], rem_address6[33];
- unsigned int local_address, local_port, state, rem_address, rem_port, inode;
-
- switch(protocol) {
- case PROC_NET_PROTOCOL_TCP:
- if(line[34] != '0' || line[35] != 'A')
- continue;
- // fall-through
-
- case PROC_NET_PROTOCOL_UDP:
- if (sscanf(line, "%*d: %X:%X %X:%X %X %*X:%*X %*X:%*X %*X %*d %*d %u",
- &local_address, &local_port, &rem_address, &rem_port, &state, &inode) != 6)
- continue;
-
- print_ipv4_address(local_address, address);
- break;
-
- case PROC_NET_PROTOCOL_TCP6:
- if(line[82] != '0' || line[83] != 'A')
- continue;
- // fall-through
-
- case PROC_NET_PROTOCOL_UDP6:
- if(sscanf(line, "%*d: %32[0-9A-Fa-f]:%X %32[0-9A-Fa-f]:%X %X %*X:%*X %*X:%*X %*X %*d %*d %u",
- local_address6, &local_port, rem_address6, &rem_port, &state, &inode) != 6)
- continue;
-
- print_ipv6_address(local_address6, address);
- break;
- }
-
- add_port_and_inode(protocol, local_port, inode, address);
- }
-
- fclose(fp);
- if (line)
- free(line);
-
- return true;
-}
-
-// ----------------------------------------------------------------------------
-typedef struct {
- bool read_tcp;
- bool read_tcp6;
- bool read_udp;
- bool read_udp6;
-} CommandLineArguments;
-
-int main(int argc, char **argv) {
- char path[FILENAME_MAX + 1];
- hashTable_key_inode_port_value = createHashTable();
-
- netdata_configured_host_prefix = getenv("NETDATA_HOST_PREFIX");
- if(!netdata_configured_host_prefix) netdata_configured_host_prefix = "";
-
- CommandLineArguments args = {.read_tcp = false, .read_tcp6 = false, .read_udp = false, .read_udp6 = false};
-
- for (int i = 1; i < argc; i++) {
- if (strcmp("tcp", argv[i]) == 0) {
- args.read_tcp = true;
- continue;
- } else if (strcmp("tcp6", argv[i]) == 0) {
- args.read_tcp6 = true;
- continue;
- } else if (strcmp("udp", argv[i]) == 0) {
- args.read_udp = true;
- continue;
- } else if (strcmp("udp6", argv[i]) == 0) {
- args.read_udp6 = true;
- continue;
- }
- }
-
- bool read_all_files = (!args.read_tcp && !args.read_tcp6 && !args.read_udp && !args.read_udp6);
-
- if (read_all_files || args.read_tcp) {
- snprintfz(path, FILENAME_MAX, "%s/proc/net/tcp", netdata_configured_host_prefix);
- read_proc_net_x(path, PROC_NET_PROTOCOL_TCP);
- }
-
- if (read_all_files || args.read_udp) {
- snprintfz(path, FILENAME_MAX, "%s/proc/net/udp", netdata_configured_host_prefix);
- read_proc_net_x(path, PROC_NET_PROTOCOL_UDP);
- }
-
- if (read_all_files || args.read_tcp6) {
- snprintfz(path, FILENAME_MAX, "%s/proc/net/tcp6", netdata_configured_host_prefix);
- read_proc_net_x(path, PROC_NET_PROTOCOL_TCP6);
- }
-
- if (read_all_files || args.read_udp6) {
- snprintfz(path, FILENAME_MAX, "%s/proc/net/udp6", netdata_configured_host_prefix);
- read_proc_net_x(path, PROC_NET_PROTOCOL_UDP6);
- }
-
- snprintfz(path, FILENAME_MAX, "%s/proc", netdata_configured_host_prefix);
- find_all_sockets_in_proc(path);
-
- freeHashTable(hashTable_key_inode_port_value);
- return 0;
-}
diff --git a/collectors/plugins.d/plugins_d.h b/collectors/plugins.d/plugins_d.h
deleted file mode 100644
index 37c70f7e3..000000000
--- a/collectors/plugins.d/plugins_d.h
+++ /dev/null
@@ -1,67 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-#ifndef NETDATA_PLUGINS_D_H
-#define NETDATA_PLUGINS_D_H 1
-
-#include "daemon/common.h"
-
-#define PLUGINSD_FILE_SUFFIX ".plugin"
-#define PLUGINSD_FILE_SUFFIX_LEN strlen(PLUGINSD_FILE_SUFFIX)
-#define PLUGINSD_CMD_MAX (FILENAME_MAX*2)
-#define PLUGINSD_STOCK_PLUGINS_DIRECTORY_PATH 0
-
-#define PLUGINSD_KEYWORD_FUNCTION_PAYLOAD "FUNCTION_PAYLOAD"
-#define PLUGINSD_KEYWORD_FUNCTION_PAYLOAD_END "FUNCTION_PAYLOAD_END"
-
-#define PLUGINSD_KEYWORD_DYNCFG_ENABLE "DYNCFG_ENABLE"
-#define PLUGINSD_KEYWORD_DYNCFG_REGISTER_MODULE "DYNCFG_REGISTER_MODULE"
-#define PLUGINSD_KEYWORD_DYNCFG_REGISTER_JOB "DYNCFG_REGISTER_JOB"
-#define PLUGINSD_KEYWORD_DYNCFG_RESET "DYNCFG_RESET"
-
-#define PLUGINSD_KEYWORD_REPORT_JOB_STATUS "REPORT_JOB_STATUS"
-#define PLUGINSD_KEYWORD_DELETE_JOB "DELETE_JOB"
-
-#define PLUGINSD_MAX_DIRECTORIES 20
-extern char *plugin_directories[PLUGINSD_MAX_DIRECTORIES];
-
-struct plugind {
- char id[CONFIG_MAX_NAME+1]; // config node id
-
- char filename[FILENAME_MAX+1]; // just the filename
- char fullfilename[FILENAME_MAX+1]; // with path
- char cmd[PLUGINSD_CMD_MAX+1]; // the command that it executes
-
- size_t successful_collections; // the number of times we have seen
- // values collected from this plugin
-
- size_t serial_failures; // the number of times the plugin started
- // without collecting values
-
- RRDHOST *host; // the host the plugin collects data for
- int update_every; // the plugin default data collection frequency
-
- struct {
- SPINLOCK spinlock;
- bool running; // do not touch this structure after setting this to 1
- bool enabled; // if this is enabled or not
- netdata_thread_t thread;
- pid_t pid;
- } unsafe;
-
- time_t started_t;
-
- const DICTIONARY_ITEM *cfg_dict_item;
- struct configurable_plugin *configuration;
-
- struct plugind *prev;
- struct plugind *next;
-};
-
-extern struct plugind *pluginsd_root;
-
-size_t pluginsd_process(RRDHOST *host, struct plugind *cd, FILE *fp_plugin_input, FILE *fp_plugin_output, int trust_durations);
-void pluginsd_process_thread_cleanup(void *ptr);
-
-size_t pluginsd_initialize_plugin_directories();
-
-#endif /* NETDATA_PLUGINS_D_H */
diff --git a/collectors/plugins.d/pluginsd_parser.c b/collectors/plugins.d/pluginsd_parser.c
deleted file mode 100644
index 3b47c6c0f..000000000
--- a/collectors/plugins.d/pluginsd_parser.c
+++ /dev/null
@@ -1,3208 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-#include "pluginsd_parser.h"
-
-#define LOG_FUNCTIONS false
-
-#define SERVING_STREAMING(parser) ((parser)->repertoire == PARSER_INIT_STREAMING)
-#define SERVING_PLUGINSD(parser) ((parser)->repertoire == PARSER_INIT_PLUGINSD)
-
-static ssize_t send_to_plugin(const char *txt, void *data) {
- PARSER *parser = data;
-
- if(!txt || !*txt)
- return 0;
-
-#ifdef ENABLE_H2O
- if(parser->h2o_ctx)
- return h2o_stream_write(parser->h2o_ctx, txt, strlen(txt));
-#endif
-
- errno = 0;
- spinlock_lock(&parser->writer.spinlock);
- ssize_t bytes = -1;
-
-#ifdef ENABLE_HTTPS
- NETDATA_SSL *ssl = parser->ssl_output;
- if(ssl) {
-
- if(SSL_connection(ssl))
- bytes = netdata_ssl_write(ssl, (void *) txt, strlen(txt));
-
- else
- netdata_log_error("PLUGINSD: cannot send command (SSL)");
-
- spinlock_unlock(&parser->writer.spinlock);
- return bytes;
- }
-#endif
-
- if(parser->fp_output) {
-
- bytes = fprintf(parser->fp_output, "%s", txt);
- if(bytes <= 0) {
- netdata_log_error("PLUGINSD: cannot send command (FILE)");
- bytes = -2;
- }
- else
- fflush(parser->fp_output);
-
- spinlock_unlock(&parser->writer.spinlock);
- return bytes;
- }
-
- if(parser->fd != -1) {
- bytes = 0;
- ssize_t total = (ssize_t)strlen(txt);
- ssize_t sent;
-
- do {
- sent = write(parser->fd, &txt[bytes], total - bytes);
- if(sent <= 0) {
- netdata_log_error("PLUGINSD: cannot send command (fd)");
- spinlock_unlock(&parser->writer.spinlock);
- return -3;
- }
- bytes += sent;
- }
- while(bytes < total);
-
- spinlock_unlock(&parser->writer.spinlock);
- return (int)bytes;
- }
-
- spinlock_unlock(&parser->writer.spinlock);
- netdata_log_error("PLUGINSD: cannot send command (no output socket/pipe/file given to plugins.d parser)");
- return -4;
-}
-
-static inline RRDHOST *pluginsd_require_scope_host(PARSER *parser, const char *cmd) {
- RRDHOST *host = parser->user.host;
-
- if(unlikely(!host))
- netdata_log_error("PLUGINSD: command %s requires a host, but is not set.", cmd);
-
- return host;
-}
-
-static inline RRDSET *pluginsd_require_scope_chart(PARSER *parser, const char *cmd, const char *parent_cmd) {
- RRDSET *st = parser->user.st;
-
- if(unlikely(!st))
- netdata_log_error("PLUGINSD: command %s requires a chart defined via command %s, but is not set.", cmd, parent_cmd);
-
- return st;
-}
-
-static inline RRDSET *pluginsd_get_scope_chart(PARSER *parser) {
- return parser->user.st;
-}
-
-static inline void pluginsd_lock_rrdset_data_collection(PARSER *parser) {
- if(parser->user.st && !parser->user.v2.locked_data_collection) {
- spinlock_lock(&parser->user.st->data_collection_lock);
- parser->user.v2.locked_data_collection = true;
- }
-}
-
-static inline bool pluginsd_unlock_rrdset_data_collection(PARSER *parser) {
- if(parser->user.st && parser->user.v2.locked_data_collection) {
- spinlock_unlock(&parser->user.st->data_collection_lock);
- parser->user.v2.locked_data_collection = false;
- return true;
- }
-
- return false;
-}
-
-static inline void pluginsd_unlock_previous_scope_chart(PARSER *parser, const char *keyword, bool stale) {
- if(unlikely(pluginsd_unlock_rrdset_data_collection(parser))) {
- if(stale)
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s/' stale data collection lock found during %s; it has been unlocked",
- rrdhost_hostname(parser->user.st->rrdhost),
- rrdset_id(parser->user.st),
- keyword);
- }
-
- if(unlikely(parser->user.v2.ml_locked)) {
- ml_chart_update_end(parser->user.st);
- parser->user.v2.ml_locked = false;
-
- if(stale)
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s/' stale ML lock found during %s, it has been unlocked",
- rrdhost_hostname(parser->user.st->rrdhost),
- rrdset_id(parser->user.st),
- keyword);
- }
-}
-
-static inline void pluginsd_clear_scope_chart(PARSER *parser, const char *keyword) {
- pluginsd_unlock_previous_scope_chart(parser, keyword, true);
-
- if(parser->user.cleanup_slots && parser->user.st)
- rrdset_pluginsd_receive_unslot(parser->user.st);
-
- parser->user.st = NULL;
- parser->user.cleanup_slots = false;
-}
-
-static inline bool pluginsd_set_scope_chart(PARSER *parser, RRDSET *st, const char *keyword) {
- RRDSET *old_st = parser->user.st;
- pid_t old_collector_tid = (old_st) ? old_st->pluginsd.collector_tid : 0;
- pid_t my_collector_tid = gettid();
-
- if(unlikely(old_collector_tid)) {
- if(old_collector_tid != my_collector_tid) {
- nd_log_limit_static_global_var(erl, 1, 0);
- nd_log_limit(&erl, NDLS_COLLECTORS, NDLP_WARNING,
- "PLUGINSD: keyword %s: 'host:%s/chart:%s' is collected twice (my tid %d, other collector tid %d)",
- keyword ? keyword : "UNKNOWN",
- rrdhost_hostname(st->rrdhost), rrdset_id(st),
- my_collector_tid, old_collector_tid);
-
- return false;
- }
-
- old_st->pluginsd.collector_tid = 0;
- }
-
- st->pluginsd.collector_tid = my_collector_tid;
-
- pluginsd_clear_scope_chart(parser, keyword);
-
- st->pluginsd.pos = 0;
- parser->user.st = st;
- parser->user.cleanup_slots = false;
-
- return true;
-}
-
-static inline void pluginsd_rrddim_put_to_slot(PARSER *parser, RRDSET *st, RRDDIM *rd, ssize_t slot, bool obsolete) {
- size_t wanted_size = st->pluginsd.size;
-
- if(slot >= 1) {
- st->pluginsd.dims_with_slots = true;
- wanted_size = slot;
- }
- else {
- st->pluginsd.dims_with_slots = false;
- wanted_size = dictionary_entries(st->rrddim_root_index);
- }
-
- if(wanted_size > st->pluginsd.size) {
- st->pluginsd.prd_array = reallocz(st->pluginsd.prd_array, wanted_size * sizeof(struct pluginsd_rrddim));
-
- // initialize the empty slots
- for(ssize_t i = (ssize_t) wanted_size - 1; i >= (ssize_t) st->pluginsd.size; i--) {
- st->pluginsd.prd_array[i].rda = NULL;
- st->pluginsd.prd_array[i].rd = NULL;
- st->pluginsd.prd_array[i].id = NULL;
- }
-
- st->pluginsd.size = wanted_size;
- }
-
- if(st->pluginsd.dims_with_slots) {
- struct pluginsd_rrddim *prd = &st->pluginsd.prd_array[slot - 1];
-
- if(prd->rd != rd) {
- prd->rda = rrddim_find_and_acquire(st, string2str(rd->id));
- prd->rd = rrddim_acquired_to_rrddim(prd->rda);
- prd->id = string2str(prd->rd->id);
- }
-
- if(obsolete)
- parser->user.cleanup_slots = true;
- }
-}
-
-static inline RRDDIM *pluginsd_acquire_dimension(RRDHOST *host, RRDSET *st, const char *dimension, ssize_t slot, const char *cmd) {
- if (unlikely(!dimension || !*dimension)) {
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s' got a %s, without a dimension.",
- rrdhost_hostname(host), rrdset_id(st), cmd);
- return NULL;
- }
-
- if (unlikely(!st->pluginsd.size)) {
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s' got a %s, but the chart has no dimensions.",
- rrdhost_hostname(host), rrdset_id(st), cmd);
- return NULL;
- }
-
- struct pluginsd_rrddim *prd;
- RRDDIM *rd;
-
- if(likely(st->pluginsd.dims_with_slots)) {
- // caching with slots
-
- if(unlikely(slot < 1 || slot > st->pluginsd.size)) {
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s' got a %s with slot %zd, but slots in the range [1 - %u] are expected.",
- rrdhost_hostname(host), rrdset_id(st), cmd, slot, st->pluginsd.size);
- return NULL;
- }
-
- prd = &st->pluginsd.prd_array[slot - 1];
-
- rd = prd->rd;
- if(likely(rd)) {
-#ifdef NETDATA_INTERNAL_CHECKS
- if(strcmp(prd->id, dimension) != 0) {
- ssize_t t;
- for(t = 0; t < st->pluginsd.size ;t++) {
- if (strcmp(st->pluginsd.prd_array[t].id, dimension) == 0)
- break;
- }
- if(t >= st->pluginsd.size)
- t = -1;
-
- internal_fatal(true,
- "PLUGINSD: expected to find dimension '%s' on slot %zd, but found '%s', "
- "the right slot is %zd",
- dimension, slot, prd->id, t);
- }
-#endif
- return rd;
- }
- }
- else {
- // caching without slots
-
- if(unlikely(st->pluginsd.pos >= st->pluginsd.size))
- st->pluginsd.pos = 0;
-
- prd = &st->pluginsd.prd_array[st->pluginsd.pos++];
-
- rd = prd->rd;
- if(likely(rd)) {
- const char *id = prd->id;
-
- if(strcmp(id, dimension) == 0) {
- // we found it cached
- return rd;
- }
- else {
- // the cached one is not good for us
- rrddim_acquired_release(prd->rda);
- prd->rda = NULL;
- prd->rd = NULL;
- prd->id = NULL;
- }
- }
- }
-
- // we need to find the dimension and set it to prd
-
- RRDDIM_ACQUIRED *rda = rrddim_find_and_acquire(st, dimension);
- if (unlikely(!rda)) {
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s/dim:%s' got a %s but dimension does not exist.",
- rrdhost_hostname(host), rrdset_id(st), dimension, cmd);
-
- return NULL;
- }
-
- prd->rda = rda;
- prd->rd = rd = rrddim_acquired_to_rrddim(rda);
- prd->id = string2str(rd->id);
-
- return rd;
-}
-
-static inline RRDSET *pluginsd_find_chart(RRDHOST *host, const char *chart, const char *cmd) {
- if (unlikely(!chart || !*chart)) {
- netdata_log_error("PLUGINSD: 'host:%s' got a %s without a chart id.",
- rrdhost_hostname(host), cmd);
- return NULL;
- }
-
- RRDSET *st = rrdset_find(host, chart);
- if (unlikely(!st))
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s' got a %s but chart does not exist.",
- rrdhost_hostname(host), chart, cmd);
-
- return st;
-}
-
-static inline ssize_t pluginsd_parse_rrd_slot(char **words, size_t num_words) {
- ssize_t slot = -1;
- char *id = get_word(words, num_words, 1);
- if(id && id[0] == PLUGINSD_KEYWORD_SLOT[0] && id[1] == PLUGINSD_KEYWORD_SLOT[1] &&
- id[2] == PLUGINSD_KEYWORD_SLOT[2] && id[3] == PLUGINSD_KEYWORD_SLOT[3] && id[4] == ':') {
- slot = (ssize_t) str2ull_encoded(&id[5]);
- if(slot < 0) slot = 0; // to make the caller increment its idx of the words
- }
-
- return slot;
-}
-
-static inline void pluginsd_rrdset_cache_put_to_slot(PARSER *parser, RRDSET *st, ssize_t slot, bool obsolete) {
- // clean possible old cached data
- rrdset_pluginsd_receive_unslot(st);
-
- if(unlikely(slot < 1 || slot >= INT32_MAX))
- return;
-
- RRDHOST *host = st->rrdhost;
-
- if(unlikely((size_t)slot > host->rrdpush.receive.pluginsd_chart_slots.size)) {
- spinlock_lock(&host->rrdpush.receive.pluginsd_chart_slots.spinlock);
- size_t old_slots = host->rrdpush.receive.pluginsd_chart_slots.size;
- size_t new_slots = (old_slots < PLUGINSD_MIN_RRDSET_POINTERS_CACHE) ? PLUGINSD_MIN_RRDSET_POINTERS_CACHE : old_slots * 2;
-
- if(new_slots < (size_t)slot)
- new_slots = slot;
-
- host->rrdpush.receive.pluginsd_chart_slots.array =
- reallocz(host->rrdpush.receive.pluginsd_chart_slots.array, new_slots * sizeof(RRDSET *));
-
- for(size_t i = old_slots; i < new_slots ;i++)
- host->rrdpush.receive.pluginsd_chart_slots.array[i] = NULL;
-
- host->rrdpush.receive.pluginsd_chart_slots.size = new_slots;
- spinlock_unlock(&host->rrdpush.receive.pluginsd_chart_slots.spinlock);
- }
-
- host->rrdpush.receive.pluginsd_chart_slots.array[slot - 1] = st;
- st->pluginsd.last_slot = (int32_t)slot - 1;
- parser->user.cleanup_slots = obsolete;
-}
-
-static inline RRDSET *pluginsd_rrdset_cache_get_from_slot(PARSER *parser, RRDHOST *host, const char *id, ssize_t slot, const char *keyword) {
- if(unlikely(slot < 1 || (size_t)slot > host->rrdpush.receive.pluginsd_chart_slots.size))
- return pluginsd_find_chart(host, id, keyword);
-
- RRDSET *st = host->rrdpush.receive.pluginsd_chart_slots.array[slot - 1];
-
- if(!st) {
- st = pluginsd_find_chart(host, id, keyword);
- if(st)
- pluginsd_rrdset_cache_put_to_slot(parser, st, slot, rrdset_flag_check(st, RRDSET_FLAG_OBSOLETE));
- }
- else {
- internal_fatal(string_strcmp(st->id, id) != 0,
- "PLUGINSD: wrong chart in slot %zd, expected '%s', found '%s'",
- slot - 1, id, string2str(st->id));
- }
-
- return st;
-}
-
-static inline PARSER_RC PLUGINSD_DISABLE_PLUGIN(PARSER *parser, const char *keyword, const char *msg) {
- parser->user.enabled = 0;
-
- if(keyword && msg) {
- nd_log_limit_static_global_var(erl, 1, 0);
- nd_log_limit(&erl, NDLS_COLLECTORS, NDLP_INFO,
- "PLUGINSD: keyword %s: %s", keyword, msg);
- }
-
- return PARSER_RC_ERROR;
-}
-
-static inline PARSER_RC pluginsd_set(char **words, size_t num_words, PARSER *parser) {
- int idx = 1;
- ssize_t slot = pluginsd_parse_rrd_slot(words, num_words);
- if(slot >= 0) idx++;
-
- char *dimension = get_word(words, num_words, idx++);
- char *value = get_word(words, num_words, idx++);
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_SET);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_SET, PLUGINSD_KEYWORD_CHART);
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDDIM *rd = pluginsd_acquire_dimension(host, st, dimension, slot, PLUGINSD_KEYWORD_SET);
- if(!rd) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- st->pluginsd.set = true;
-
- if (unlikely(rrdset_flag_check(st, RRDSET_FLAG_DEBUG)))
- netdata_log_debug(D_PLUGINSD, "PLUGINSD: 'host:%s/chart:%s/dim:%s' SET is setting value to '%s'",
- rrdhost_hostname(host), rrdset_id(st), dimension, value && *value ? value : "UNSET");
-
- if (value && *value)
- rrddim_set_by_pointer(st, rd, str2ll_encoded(value));
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_begin(char **words, size_t num_words, PARSER *parser) {
- int idx = 1;
- ssize_t slot = pluginsd_parse_rrd_slot(words, num_words);
- if(slot >= 0) idx++;
-
- char *id = get_word(words, num_words, idx++);
- char *microseconds_txt = get_word(words, num_words, idx++);
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_BEGIN);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_rrdset_cache_get_from_slot(parser, host, id, slot, PLUGINSD_KEYWORD_BEGIN);
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- if(!pluginsd_set_scope_chart(parser, st, PLUGINSD_KEYWORD_BEGIN))
- return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- usec_t microseconds = 0;
- if (microseconds_txt && *microseconds_txt) {
- long long t = str2ll(microseconds_txt, NULL);
- if(t >= 0)
- microseconds = t;
- }
-
-#ifdef NETDATA_LOG_REPLICATION_REQUESTS
- if(st->replay.log_next_data_collection) {
- st->replay.log_next_data_collection = false;
-
- internal_error(true,
- "REPLAY: 'host:%s/chart:%s' first BEGIN after replication, last collected %llu, last updated %llu, microseconds %llu",
- rrdhost_hostname(host), rrdset_id(st),
- st->last_collected_time.tv_sec * USEC_PER_SEC + st->last_collected_time.tv_usec,
- st->last_updated.tv_sec * USEC_PER_SEC + st->last_updated.tv_usec,
- microseconds
- );
- }
-#endif
-
- if (likely(st->counter_done)) {
- if (likely(microseconds)) {
- if (parser->user.trust_durations)
- rrdset_next_usec_unfiltered(st, microseconds);
- else
- rrdset_next_usec(st, microseconds);
- }
- else
- rrdset_next(st);
- }
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_end(char **words, size_t num_words, PARSER *parser) {
- char *tv_sec = get_word(words, num_words, 1);
- char *tv_usec = get_word(words, num_words, 2);
- char *pending_rrdset_next = get_word(words, num_words, 3);
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_END);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_END, PLUGINSD_KEYWORD_BEGIN);
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- if (unlikely(rrdset_flag_check(st, RRDSET_FLAG_DEBUG)))
- netdata_log_debug(D_PLUGINSD, "requested an END on chart '%s'", rrdset_id(st));
-
- pluginsd_clear_scope_chart(parser, PLUGINSD_KEYWORD_END);
- parser->user.data_collections_count++;
-
- struct timeval tv = {
- .tv_sec = (tv_sec && *tv_sec) ? str2ll(tv_sec, NULL) : 0,
- .tv_usec = (tv_usec && *tv_usec) ? str2ll(tv_usec, NULL) : 0
- };
-
- if(!tv.tv_sec)
- now_realtime_timeval(&tv);
-
- rrdset_timed_done(st, tv, pending_rrdset_next && *pending_rrdset_next ? true : false);
-
- return PARSER_RC_OK;
-}
-
-static void pluginsd_host_define_cleanup(PARSER *parser) {
- string_freez(parser->user.host_define.hostname);
- rrdlabels_destroy(parser->user.host_define.rrdlabels);
-
- parser->user.host_define.hostname = NULL;
- parser->user.host_define.rrdlabels = NULL;
- parser->user.host_define.parsing_host = false;
-}
-
-static inline bool pluginsd_validate_machine_guid(const char *guid, uuid_t *uuid, char *output) {
- if(uuid_parse(guid, *uuid))
- return false;
-
- uuid_unparse_lower(*uuid, output);
-
- return true;
-}
-
-static inline PARSER_RC pluginsd_host_define(char **words, size_t num_words, PARSER *parser) {
- char *guid = get_word(words, num_words, 1);
- char *hostname = get_word(words, num_words, 2);
-
- if(unlikely(!guid || !*guid || !hostname || !*hostname))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_HOST_DEFINE, "missing parameters");
-
- if(unlikely(parser->user.host_define.parsing_host))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_HOST_DEFINE,
- "another host definition is already open - did you send " PLUGINSD_KEYWORD_HOST_DEFINE_END "?");
-
- if(!pluginsd_validate_machine_guid(guid, &parser->user.host_define.machine_guid, parser->user.host_define.machine_guid_str))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_HOST_DEFINE, "cannot parse MACHINE_GUID - is it a valid UUID?");
-
- parser->user.host_define.hostname = string_strdupz(hostname);
- parser->user.host_define.rrdlabels = rrdlabels_create();
- parser->user.host_define.parsing_host = true;
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_host_dictionary(char **words, size_t num_words, PARSER *parser, RRDLABELS *labels, const char *keyword) {
- char *name = get_word(words, num_words, 1);
- char *value = get_word(words, num_words, 2);
-
- if(!name || !*name || !value)
- return PLUGINSD_DISABLE_PLUGIN(parser, keyword, "missing parameters");
-
- if(!parser->user.host_define.parsing_host || !labels)
- return PLUGINSD_DISABLE_PLUGIN(parser, keyword, "host is not defined, send " PLUGINSD_KEYWORD_HOST_DEFINE " before this");
-
- rrdlabels_add(labels, name, value, RRDLABEL_SRC_CONFIG);
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_host_labels(char **words, size_t num_words, PARSER *parser) {
- return pluginsd_host_dictionary(words, num_words, parser,
- parser->user.host_define.rrdlabels,
- PLUGINSD_KEYWORD_HOST_LABEL);
-}
-
-static inline PARSER_RC pluginsd_host_define_end(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser) {
- if(!parser->user.host_define.parsing_host)
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_HOST_DEFINE_END, "missing initialization, send " PLUGINSD_KEYWORD_HOST_DEFINE " before this");
-
- RRDHOST *host = rrdhost_find_or_create(
- string2str(parser->user.host_define.hostname),
- string2str(parser->user.host_define.hostname),
- parser->user.host_define.machine_guid_str,
- "Netdata Virtual Host 1.0",
- netdata_configured_timezone,
- netdata_configured_abbrev_timezone,
- netdata_configured_utc_offset,
- NULL,
- program_name,
- program_version,
- default_rrd_update_every,
- default_rrd_history_entries,
- default_rrd_memory_mode,
- default_health_enabled,
- default_rrdpush_enabled,
- default_rrdpush_destination,
- default_rrdpush_api_key,
- default_rrdpush_send_charts_matching,
- default_rrdpush_enable_replication,
- default_rrdpush_seconds_to_replicate,
- default_rrdpush_replication_step,
- rrdhost_labels_to_system_info(parser->user.host_define.rrdlabels),
- false);
-
- rrdhost_option_set(host, RRDHOST_OPTION_VIRTUAL_HOST);
-
- if(host->rrdlabels) {
- rrdlabels_migrate_to_these(host->rrdlabels, parser->user.host_define.rrdlabels);
- }
- else {
- host->rrdlabels = parser->user.host_define.rrdlabels;
- parser->user.host_define.rrdlabels = NULL;
- }
-
- pluginsd_host_define_cleanup(parser);
-
- parser->user.host = host;
- pluginsd_clear_scope_chart(parser, PLUGINSD_KEYWORD_HOST_DEFINE_END);
-
- rrdhost_flag_clear(host, RRDHOST_FLAG_ORPHAN);
- rrdcontext_host_child_connected(host);
- schedule_node_info_update(host);
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_host(char **words, size_t num_words, PARSER *parser) {
- char *guid = get_word(words, num_words, 1);
-
- if(!guid || !*guid || strcmp(guid, "localhost") == 0) {
- parser->user.host = localhost;
- return PARSER_RC_OK;
- }
-
- uuid_t uuid;
- char uuid_str[UUID_STR_LEN];
- if(!pluginsd_validate_machine_guid(guid, &uuid, uuid_str))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_HOST, "cannot parse MACHINE_GUID - is it a valid UUID?");
-
- RRDHOST *host = rrdhost_find_by_guid(uuid_str);
- if(unlikely(!host))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_HOST, "cannot find a host with this machine guid - have you created it?");
-
- parser->user.host = host;
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_chart(char **words, size_t num_words, PARSER *parser) {
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_CHART);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- int idx = 1;
- ssize_t slot = pluginsd_parse_rrd_slot(words, num_words);
- if(slot >= 0) idx++;
-
- char *type = get_word(words, num_words, idx++);
- char *name = get_word(words, num_words, idx++);
- char *title = get_word(words, num_words, idx++);
- char *units = get_word(words, num_words, idx++);
- char *family = get_word(words, num_words, idx++);
- char *context = get_word(words, num_words, idx++);
- char *chart = get_word(words, num_words, idx++);
- char *priority_s = get_word(words, num_words, idx++);
- char *update_every_s = get_word(words, num_words, idx++);
- char *options = get_word(words, num_words, idx++);
- char *plugin = get_word(words, num_words, idx++);
- char *module = get_word(words, num_words, idx++);
-
- // parse the id from type
- char *id = NULL;
- if (likely(type && (id = strchr(type, '.')))) {
- *id = '\0';
- id++;
- }
-
- // make sure we have the required variables
- if (unlikely((!type || !*type || !id || !*id)))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_CHART, "missing parameters");
-
- // parse the name, and make sure it does not include 'type.'
- if (unlikely(name && *name)) {
- // when data are streamed from child nodes
- // name will be type.name
- // so, we have to remove 'type.' from name too
- size_t len = strlen(type);
- if (strncmp(type, name, len) == 0 && name[len] == '.')
- name = &name[len + 1];
-
- // if the name is the same with the id,
- // or is just 'NULL', clear it.
- if (unlikely(strcmp(name, id) == 0 || strcasecmp(name, "NULL") == 0 || strcasecmp(name, "(NULL)") == 0))
- name = NULL;
- }
-
- int priority = 1000;
- if (likely(priority_s && *priority_s))
- priority = str2i(priority_s);
-
- int update_every = parser->user.cd->update_every;
- if (likely(update_every_s && *update_every_s))
- update_every = str2i(update_every_s);
- if (unlikely(!update_every))
- update_every = parser->user.cd->update_every;
-
- RRDSET_TYPE chart_type = RRDSET_TYPE_LINE;
- if (unlikely(chart))
- chart_type = rrdset_type_id(chart);
-
- if (unlikely(name && !*name))
- name = NULL;
- if (unlikely(family && !*family))
- family = NULL;
- if (unlikely(context && !*context))
- context = NULL;
- if (unlikely(!title))
- title = "";
- if (unlikely(!units))
- units = "unknown";
-
- netdata_log_debug(
- D_PLUGINSD,
- "creating chart type='%s', id='%s', name='%s', family='%s', context='%s', chart='%s', priority=%d, update_every=%d",
- type, id, name ? name : "", family ? family : "", context ? context : "", rrdset_type_name(chart_type),
- priority, update_every);
-
- RRDSET *st = NULL;
-
- st = rrdset_create(
- host, type, id, name, family, context, title, units,
- (plugin && *plugin) ? plugin : parser->user.cd->filename,
- module, priority, update_every,
- chart_type);
-
- bool obsolete = false;
- if (likely(st)) {
- if (options && *options) {
- if (strstr(options, "obsolete")) {
- rrdset_is_obsolete___safe_from_collector_thread(st);
- obsolete = true;
- }
- else
- rrdset_isnot_obsolete___safe_from_collector_thread(st);
-
- if (strstr(options, "detail"))
- rrdset_flag_set(st, RRDSET_FLAG_DETAIL);
- else
- rrdset_flag_clear(st, RRDSET_FLAG_DETAIL);
-
- if (strstr(options, "hidden"))
- rrdset_flag_set(st, RRDSET_FLAG_HIDDEN);
- else
- rrdset_flag_clear(st, RRDSET_FLAG_HIDDEN);
-
- if (strstr(options, "store_first"))
- rrdset_flag_set(st, RRDSET_FLAG_STORE_FIRST);
- else
- rrdset_flag_clear(st, RRDSET_FLAG_STORE_FIRST);
- }
- else {
- rrdset_isnot_obsolete___safe_from_collector_thread(st);
- rrdset_flag_clear(st, RRDSET_FLAG_DETAIL);
- rrdset_flag_clear(st, RRDSET_FLAG_STORE_FIRST);
- }
-
- if(!pluginsd_set_scope_chart(parser, st, PLUGINSD_KEYWORD_CHART))
- return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- pluginsd_rrdset_cache_put_to_slot(parser, st, slot, obsolete);
- }
- else
- pluginsd_clear_scope_chart(parser, PLUGINSD_KEYWORD_CHART);
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_chart_definition_end(char **words, size_t num_words, PARSER *parser) {
- const char *first_entry_txt = get_word(words, num_words, 1);
- const char *last_entry_txt = get_word(words, num_words, 2);
- const char *wall_clock_time_txt = get_word(words, num_words, 3);
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_CHART_DEFINITION_END);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_CHART_DEFINITION_END, PLUGINSD_KEYWORD_CHART);
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- time_t first_entry_child = (first_entry_txt && *first_entry_txt) ? (time_t)str2ul(first_entry_txt) : 0;
- time_t last_entry_child = (last_entry_txt && *last_entry_txt) ? (time_t)str2ul(last_entry_txt) : 0;
- time_t child_wall_clock_time = (wall_clock_time_txt && *wall_clock_time_txt) ? (time_t)str2ul(wall_clock_time_txt) : now_realtime_sec();
-
- bool ok = true;
- if(!rrdset_flag_check(st, RRDSET_FLAG_RECEIVER_REPLICATION_IN_PROGRESS)) {
-
-#ifdef NETDATA_LOG_REPLICATION_REQUESTS
- st->replay.start_streaming = false;
- st->replay.after = 0;
- st->replay.before = 0;
-#endif
-
- rrdset_flag_set(st, RRDSET_FLAG_RECEIVER_REPLICATION_IN_PROGRESS);
- rrdset_flag_clear(st, RRDSET_FLAG_RECEIVER_REPLICATION_FINISHED);
- rrdhost_receiver_replicating_charts_plus_one(st->rrdhost);
-
- ok = replicate_chart_request(send_to_plugin, parser, host, st,
- first_entry_child, last_entry_child, child_wall_clock_time,
- 0, 0);
- }
-#ifdef NETDATA_LOG_REPLICATION_REQUESTS
- else {
- internal_error(true, "REPLAY: 'host:%s/chart:%s' not sending duplicate replication request",
- rrdhost_hostname(st->rrdhost), rrdset_id(st));
- }
-#endif
-
- return ok ? PARSER_RC_OK : PARSER_RC_ERROR;
-}
-
-static inline PARSER_RC pluginsd_dimension(char **words, size_t num_words, PARSER *parser) {
- int idx = 1;
- ssize_t slot = pluginsd_parse_rrd_slot(words, num_words);
- if(slot >= 0) idx++;
-
- char *id = get_word(words, num_words, idx++);
- char *name = get_word(words, num_words, idx++);
- char *algorithm = get_word(words, num_words, idx++);
- char *multiplier_s = get_word(words, num_words, idx++);
- char *divisor_s = get_word(words, num_words, idx++);
- char *options = get_word(words, num_words, idx++);
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_DIMENSION);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_DIMENSION, PLUGINSD_KEYWORD_CHART);
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- if (unlikely(!id))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DIMENSION, "missing dimension id");
-
- long multiplier = 1;
- if (multiplier_s && *multiplier_s) {
- multiplier = str2ll_encoded(multiplier_s);
- if (unlikely(!multiplier))
- multiplier = 1;
- }
-
- long divisor = 1;
- if (likely(divisor_s && *divisor_s)) {
- divisor = str2ll_encoded(divisor_s);
- if (unlikely(!divisor))
- divisor = 1;
- }
-
- if (unlikely(!algorithm || !*algorithm))
- algorithm = "absolute";
-
- if (unlikely(st && rrdset_flag_check(st, RRDSET_FLAG_DEBUG)))
- netdata_log_debug(
- D_PLUGINSD,
- "creating dimension in chart %s, id='%s', name='%s', algorithm='%s', multiplier=%ld, divisor=%ld, hidden='%s'",
- rrdset_id(st), id, name ? name : "", rrd_algorithm_name(rrd_algorithm_id(algorithm)), multiplier, divisor,
- options ? options : "");
-
- RRDDIM *rd = rrddim_add(st, id, name, multiplier, divisor, rrd_algorithm_id(algorithm));
- int unhide_dimension = 1;
-
- rrddim_option_clear(rd, RRDDIM_OPTION_DONT_DETECT_RESETS_OR_OVERFLOWS);
- bool obsolete = false;
- if (options && *options) {
- if (strstr(options, "obsolete") != NULL) {
- obsolete = true;
- rrddim_is_obsolete___safe_from_collector_thread(st, rd);
- }
- else
- rrddim_isnot_obsolete___safe_from_collector_thread(st, rd);
-
- unhide_dimension = !strstr(options, "hidden");
-
- if (strstr(options, "noreset") != NULL)
- rrddim_option_set(rd, RRDDIM_OPTION_DONT_DETECT_RESETS_OR_OVERFLOWS);
- if (strstr(options, "nooverflow") != NULL)
- rrddim_option_set(rd, RRDDIM_OPTION_DONT_DETECT_RESETS_OR_OVERFLOWS);
- }
- else
- rrddim_isnot_obsolete___safe_from_collector_thread(st, rd);
-
- bool should_update_dimension = false;
-
- if (likely(unhide_dimension)) {
- rrddim_option_clear(rd, RRDDIM_OPTION_HIDDEN);
- should_update_dimension = rrddim_flag_check(rd, RRDDIM_FLAG_META_HIDDEN);
- }
- else {
- rrddim_option_set(rd, RRDDIM_OPTION_HIDDEN);
- should_update_dimension = !rrddim_flag_check(rd, RRDDIM_FLAG_META_HIDDEN);
- }
-
- if (should_update_dimension) {
- rrddim_flag_set(rd, RRDDIM_FLAG_METADATA_UPDATE);
- rrdhost_flag_set(rd->rrdset->rrdhost, RRDHOST_FLAG_METADATA_UPDATE);
- }
-
- pluginsd_rrddim_put_to_slot(parser, st, rd, slot, obsolete);
-
- return PARSER_RC_OK;
-}
-
-// ----------------------------------------------------------------------------
-// execution of functions
-
-struct inflight_function {
- int code;
- int timeout;
- STRING *function;
- BUFFER *result_body_wb;
- rrd_function_result_callback_t result_cb;
- void *result_cb_data;
- usec_t timeout_ut;
- usec_t started_ut;
- usec_t sent_ut;
- const char *payload;
- PARSER *parser;
- bool virtual;
-};
-
-static void inflight_functions_insert_callback(const DICTIONARY_ITEM *item, void *func, void *parser_ptr) {
- struct inflight_function *pf = func;
-
- PARSER *parser = parser_ptr;
-
- // leave this code as default, so that when the dictionary is destroyed this will be sent back to the caller
- pf->code = HTTP_RESP_GATEWAY_TIMEOUT;
-
- const char *transaction = dictionary_acquired_item_name(item);
-
- char buffer[2048 + 1];
- snprintfz(buffer, sizeof(buffer) - 1, "%s %s %d \"%s\"\n",
- pf->payload ? "FUNCTION_PAYLOAD" : "FUNCTION",
- transaction,
- pf->timeout,
- string2str(pf->function));
-
- // send the command to the plugin
- ssize_t ret = send_to_plugin(buffer, parser);
-
- pf->sent_ut = now_realtime_usec();
-
- if(ret < 0) {
- netdata_log_error("FUNCTION '%s': failed to send it to the plugin, error %zd", string2str(pf->function), ret);
- rrd_call_function_error(pf->result_body_wb, "Failed to communicate with collector", HTTP_RESP_SERVICE_UNAVAILABLE);
- }
- else {
- internal_error(LOG_FUNCTIONS,
- "FUNCTION '%s' with transaction '%s' sent to collector (%zd bytes, in %"PRIu64" usec)",
- string2str(pf->function), dictionary_acquired_item_name(item), ret,
- pf->sent_ut - pf->started_ut);
- }
-
- if (!pf->payload)
- return;
-
- // send the payload to the plugin
- ret = send_to_plugin(pf->payload, parser);
-
- if(ret < 0) {
- netdata_log_error("FUNCTION_PAYLOAD '%s': failed to send function to plugin, error %zd", string2str(pf->function), ret);
- rrd_call_function_error(pf->result_body_wb, "Failed to communicate with collector", HTTP_RESP_SERVICE_UNAVAILABLE);
- }
- else {
- internal_error(LOG_FUNCTIONS,
- "FUNCTION_PAYLOAD '%s' with transaction '%s' sent to collector (%zd bytes, in %"PRIu64" usec)",
- string2str(pf->function), dictionary_acquired_item_name(item), ret,
- pf->sent_ut - pf->started_ut);
- }
-
- send_to_plugin("\nFUNCTION_PAYLOAD_END\n", parser);
-}
-
-static bool inflight_functions_conflict_callback(const DICTIONARY_ITEM *item __maybe_unused, void *func __maybe_unused, void *new_func, void *parser_ptr __maybe_unused) {
- struct inflight_function *pf = new_func;
-
- netdata_log_error("PLUGINSD_PARSER: duplicate UUID on pending function '%s' detected. Ignoring the second one.", string2str(pf->function));
- pf->code = rrd_call_function_error(pf->result_body_wb, "This request is already in progress", HTTP_RESP_BAD_REQUEST);
- pf->result_cb(pf->result_body_wb, pf->code, pf->result_cb_data);
- string_freez(pf->function);
-
- return false;
-}
-
-void delete_job_finalize(struct parser *parser __maybe_unused, struct configurable_plugin *plug, const char *fnc_sig, int code) {
- if (code != DYNCFG_VFNC_RET_CFG_ACCEPTED)
- return;
-
- char *params_local = strdupz(fnc_sig);
- char *words[DYNCFG_MAX_WORDS];
- size_t words_c = quoted_strings_splitter(params_local, words, DYNCFG_MAX_WORDS, isspace_map_pluginsd);
-
- if (words_c != 3) {
- netdata_log_error("PLUGINSD_PARSER: invalid number of parameters for delete_job");
- freez(params_local);
- return;
- }
-
- const char *module = words[1];
- const char *job = words[2];
-
- delete_job(plug, module, job);
-
- unlink_job(plug->name, module, job);
-
- rrdpush_send_job_deleted(localhost, plug->name, module, job);
-
- freez(params_local);
-}
-
-void set_job_finalize(struct parser *parser __maybe_unused, struct configurable_plugin *plug __maybe_unused, const char *fnc_sig, int code) {
- if (code != DYNCFG_VFNC_RET_CFG_ACCEPTED)
- return;
-
- char *params_local = strdupz(fnc_sig);
- char *words[DYNCFG_MAX_WORDS];
- size_t words_c = quoted_strings_splitter(params_local, words, DYNCFG_MAX_WORDS, isspace_map_pluginsd);
-
- if (words_c != 3) {
- netdata_log_error("PLUGINSD_PARSER: invalid number of parameters for set_job_config");
- freez(params_local);
- return;
- }
-
- const char *module_name = get_word(words, words_c, 1);
- const char *job_name = get_word(words, words_c, 2);
-
- if (register_job(parser->user.host->configurable_plugins, parser->user.cd->configuration->name, module_name, job_name, JOB_TYPE_USER, JOB_FLG_USER_CREATED, 1)) {
- freez(params_local);
- return;
- }
-
- // only send this if it is not existing already (register_job cares for that)
- rrdpush_send_dyncfg_reg_job(localhost, parser->user.cd->configuration->name, module_name, job_name, JOB_TYPE_USER, JOB_FLG_USER_CREATED);
-
- freez(params_local);
-}
-
-static void inflight_functions_delete_callback(const DICTIONARY_ITEM *item __maybe_unused, void *func, void *parser_ptr) {
- struct inflight_function *pf = func;
- struct parser *parser = (struct parser *)parser_ptr;
-
- internal_error(LOG_FUNCTIONS,
- "FUNCTION '%s' result of transaction '%s' received from collector (%zu bytes, request %"PRIu64" usec, response %"PRIu64" usec)",
- string2str(pf->function), dictionary_acquired_item_name(item),
- buffer_strlen(pf->result_body_wb), pf->sent_ut - pf->started_ut, now_realtime_usec() - pf->sent_ut);
-
- if (pf->virtual && SERVING_PLUGINSD(parser)) {
- if (pf->payload) {
- if (strncmp(string2str(pf->function), FUNCTION_NAME_SET_JOB_CONFIG, strlen(FUNCTION_NAME_SET_JOB_CONFIG)) == 0)
- set_job_finalize(parser, parser->user.cd->configuration, string2str(pf->function), pf->code);
- dyn_conf_store_config(string2str(pf->function), pf->payload, parser->user.cd->configuration);
- } else if (strncmp(string2str(pf->function), FUNCTION_NAME_DELETE_JOB, strlen(FUNCTION_NAME_DELETE_JOB)) == 0) {
- delete_job_finalize(parser, parser->user.cd->configuration, string2str(pf->function), pf->code);
- }
- }
-
- pf->result_cb(pf->result_body_wb, pf->code, pf->result_cb_data);
-
- string_freez(pf->function);
- freez((void *)pf->payload);
-}
-
-void inflight_functions_init(PARSER *parser) {
- parser->inflight.functions = dictionary_create_advanced(DICT_OPTION_DONT_OVERWRITE_VALUE, &dictionary_stats_category_functions, 0);
- dictionary_register_insert_callback(parser->inflight.functions, inflight_functions_insert_callback, parser);
- dictionary_register_delete_callback(parser->inflight.functions, inflight_functions_delete_callback, parser);
- dictionary_register_conflict_callback(parser->inflight.functions, inflight_functions_conflict_callback, parser);
-}
-
-static void inflight_functions_garbage_collect(PARSER *parser, usec_t now) {
- parser->inflight.smaller_timeout = 0;
- struct inflight_function *pf;
- dfe_start_write(parser->inflight.functions, pf) {
- if (pf->timeout_ut < now) {
- internal_error(true,
- "FUNCTION '%s' removing expired transaction '%s', after %"PRIu64" usec.",
- string2str(pf->function), pf_dfe.name, now - pf->started_ut);
-
- if(!buffer_strlen(pf->result_body_wb) || pf->code == HTTP_RESP_OK)
- pf->code = rrd_call_function_error(pf->result_body_wb,
- "Timeout waiting for collector response.",
- HTTP_RESP_GATEWAY_TIMEOUT);
-
- dictionary_del(parser->inflight.functions, pf_dfe.name);
- }
-
- else if(!parser->inflight.smaller_timeout || pf->timeout_ut < parser->inflight.smaller_timeout)
- parser->inflight.smaller_timeout = pf->timeout_ut;
- }
- dfe_done(pf);
-}
-
-void pluginsd_function_cancel(void *data) {
- struct inflight_function *look_for = data, *t;
-
- bool sent = false;
- dfe_start_read(look_for->parser->inflight.functions, t) {
- if(look_for == t) {
- const char *transaction = t_dfe.name;
-
- internal_error(true, "PLUGINSD: sending function cancellation to plugin for transaction '%s'", transaction);
-
- char buffer[2048 + 1];
- snprintfz(buffer, sizeof(buffer) - 1, "%s %s\n",
- PLUGINSD_KEYWORD_FUNCTION_CANCEL,
- transaction);
-
- // send the command to the plugin
- ssize_t ret = send_to_plugin(buffer, t->parser);
- if(ret < 0)
- sent = true;
-
- break;
- }
- }
- dfe_done(t);
-
- if(sent <= 0)
- nd_log(NDLS_DAEMON, NDLP_DEBUG,
- "PLUGINSD: FUNCTION_CANCEL request didn't match any pending function requests in pluginsd.d.");
-}
-
-// this is the function that is called from
-// rrd_call_function_and_wait() and rrd_call_function_async()
-static int pluginsd_function_execute_cb(BUFFER *result_body_wb, int timeout, const char *function,
- void *execute_cb_data,
- rrd_function_result_callback_t result_cb, void *result_cb_data,
- rrd_function_is_cancelled_cb_t is_cancelled_cb __maybe_unused,
- void *is_cancelled_cb_data __maybe_unused,
- rrd_function_register_canceller_cb_t register_canceller_cb,
- void *register_canceller_db_data) {
- PARSER *parser = execute_cb_data;
-
- usec_t now = now_realtime_usec();
-
- struct inflight_function tmp = {
- .started_ut = now,
- .timeout_ut = now + timeout * USEC_PER_SEC + RRDFUNCTIONS_TIMEOUT_EXTENSION_UT,
- .result_body_wb = result_body_wb,
- .timeout = timeout,
- .function = string_strdupz(function),
- .result_cb = result_cb,
- .result_cb_data = result_cb_data,
- .payload = NULL,
- .parser = parser,
- };
-
- uuid_t uuid;
- uuid_generate_random(uuid);
-
- char transaction[UUID_STR_LEN];
- uuid_unparse_lower(uuid, transaction);
-
- dictionary_write_lock(parser->inflight.functions);
-
- // if there is any error, our dictionary callbacks will call the caller callback to notify
- // the caller about the error - no need for error handling here.
- void *t = dictionary_set(parser->inflight.functions, transaction, &tmp, sizeof(struct inflight_function));
- if(register_canceller_cb)
- register_canceller_cb(register_canceller_db_data, pluginsd_function_cancel, t);
-
- if(!parser->inflight.smaller_timeout || tmp.timeout_ut < parser->inflight.smaller_timeout)
- parser->inflight.smaller_timeout = tmp.timeout_ut;
-
- // garbage collect stale inflight functions
- if(parser->inflight.smaller_timeout < now)
- inflight_functions_garbage_collect(parser, now);
-
- dictionary_write_unlock(parser->inflight.functions);
-
- return HTTP_RESP_OK;
-}
-
-static inline PARSER_RC pluginsd_function(char **words, size_t num_words, PARSER *parser) {
- // a plugin or a child is registering a function
-
- bool global = false;
- size_t i = 1;
- if(num_words >= 2 && strcmp(get_word(words, num_words, 1), "GLOBAL") == 0) {
- i++;
- global = true;
- }
-
- char *name = get_word(words, num_words, i++);
- char *timeout_s = get_word(words, num_words, i++);
- char *help = get_word(words, num_words, i++);
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_FUNCTION);
- if(!host) return PARSER_RC_ERROR;
-
- RRDSET *st = (global)? NULL: pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_FUNCTION, PLUGINSD_KEYWORD_CHART);
- if(!st) global = true;
-
- if (unlikely(!timeout_s || !name || !help || (!global && !st))) {
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s' got a FUNCTION, without providing the required data (global = '%s', name = '%s', timeout = '%s', help = '%s'). Ignoring it.",
- rrdhost_hostname(host),
- st?rrdset_id(st):"(unset)",
- global?"yes":"no",
- name?name:"(unset)",
- timeout_s?timeout_s:"(unset)",
- help?help:"(unset)"
- );
- return PARSER_RC_ERROR;
- }
-
- int timeout = PLUGINS_FUNCTIONS_TIMEOUT_DEFAULT;
- if (timeout_s && *timeout_s) {
- timeout = str2i(timeout_s);
- if (unlikely(timeout <= 0))
- timeout = PLUGINS_FUNCTIONS_TIMEOUT_DEFAULT;
- }
-
- rrd_function_add(host, st, name, timeout, help, false, pluginsd_function_execute_cb, parser);
-
- parser->user.data_collections_count++;
-
- return PARSER_RC_OK;
-}
-
-static void pluginsd_function_result_end(struct parser *parser, void *action_data) {
- STRING *key = action_data;
- if(key)
- dictionary_del(parser->inflight.functions, string2str(key));
- string_freez(key);
-
- parser->user.data_collections_count++;
-}
-
-static inline PARSER_RC pluginsd_function_result_begin(char **words, size_t num_words, PARSER *parser) {
- char *key = get_word(words, num_words, 1);
- char *status = get_word(words, num_words, 2);
- char *format = get_word(words, num_words, 3);
- char *expires = get_word(words, num_words, 4);
-
- if (unlikely(!key || !*key || !status || !*status || !format || !*format || !expires || !*expires)) {
- netdata_log_error("got a " PLUGINSD_KEYWORD_FUNCTION_RESULT_BEGIN " without providing the required data (key = '%s', status = '%s', format = '%s', expires = '%s')."
- , key ? key : "(unset)"
- , status ? status : "(unset)"
- , format ? format : "(unset)"
- , expires ? expires : "(unset)"
- );
- }
-
- int code = (status && *status) ? str2i(status) : 0;
- if (code <= 0)
- code = HTTP_RESP_BACKEND_RESPONSE_INVALID;
-
- time_t expiration = (expires && *expires) ? str2l(expires) : 0;
-
- struct inflight_function *pf = NULL;
-
- if(key && *key)
- pf = (struct inflight_function *)dictionary_get(parser->inflight.functions, key);
-
- if(!pf) {
- netdata_log_error("got a " PLUGINSD_KEYWORD_FUNCTION_RESULT_BEGIN " for transaction '%s', but the transaction is not found.", key?key:"(unset)");
- }
- else {
- if(format && *format)
- pf->result_body_wb->content_type = functions_format_to_content_type(format);
-
- pf->code = code;
-
- pf->result_body_wb->expires = expiration;
- if(expiration <= now_realtime_sec())
- buffer_no_cacheable(pf->result_body_wb);
- else
- buffer_cacheable(pf->result_body_wb);
- }
-
- parser->defer.response = (pf) ? pf->result_body_wb : NULL;
- parser->defer.end_keyword = PLUGINSD_KEYWORD_FUNCTION_RESULT_END;
- parser->defer.action = pluginsd_function_result_end;
- parser->defer.action_data = string_strdupz(key); // it is ok is key is NULL
- parser->flags |= PARSER_DEFER_UNTIL_KEYWORD;
-
- return PARSER_RC_OK;
-}
-
-// ----------------------------------------------------------------------------
-
-static inline PARSER_RC pluginsd_variable(char **words, size_t num_words, PARSER *parser) {
- char *name = get_word(words, num_words, 1);
- char *value = get_word(words, num_words, 2);
- NETDATA_DOUBLE v;
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_VARIABLE);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_get_scope_chart(parser);
-
- int global = (st) ? 0 : 1;
-
- if (name && *name) {
- if ((strcmp(name, "GLOBAL") == 0 || strcmp(name, "HOST") == 0)) {
- global = 1;
- name = get_word(words, num_words, 2);
- value = get_word(words, num_words, 3);
- } else if ((strcmp(name, "LOCAL") == 0 || strcmp(name, "CHART") == 0)) {
- global = 0;
- name = get_word(words, num_words, 2);
- value = get_word(words, num_words, 3);
- }
- }
-
- if (unlikely(!name || !*name))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_VARIABLE, "missing variable name");
-
- if (unlikely(!value || !*value))
- value = NULL;
-
- if (unlikely(!value)) {
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s' cannot set %s VARIABLE '%s' to an empty value",
- rrdhost_hostname(host),
- st ? rrdset_id(st):"UNSET",
- (global) ? "HOST" : "CHART",
- name);
- return PARSER_RC_OK;
- }
-
- if (!global && !st)
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_VARIABLE, "no chart is defined and no GLOBAL is given");
-
- char *endptr = NULL;
- v = (NETDATA_DOUBLE) str2ndd_encoded(value, &endptr);
- if (unlikely(endptr && *endptr)) {
- if (endptr == value)
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s' the value '%s' of VARIABLE '%s' cannot be parsed as a number",
- rrdhost_hostname(host),
- st ? rrdset_id(st):"UNSET",
- value,
- name);
- else
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s' the value '%s' of VARIABLE '%s' has leftovers: '%s'",
- rrdhost_hostname(host),
- st ? rrdset_id(st):"UNSET",
- value,
- name,
- endptr);
- }
-
- if (global) {
- const RRDVAR_ACQUIRED *rva = rrdvar_custom_host_variable_add_and_acquire(host, name);
- if (rva) {
- rrdvar_custom_host_variable_set(host, rva, v);
- rrdvar_custom_host_variable_release(host, rva);
- }
- else
- netdata_log_error("PLUGINSD: 'host:%s' cannot find/create HOST VARIABLE '%s'",
- rrdhost_hostname(host),
- name);
- } else {
- const RRDSETVAR_ACQUIRED *rsa = rrdsetvar_custom_chart_variable_add_and_acquire(st, name);
- if (rsa) {
- rrdsetvar_custom_chart_variable_set(st, rsa, v);
- rrdsetvar_custom_chart_variable_release(st, rsa);
- }
- else
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s' cannot find/create CHART VARIABLE '%s'",
- rrdhost_hostname(host), rrdset_id(st), name);
- }
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_flush(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser) {
- netdata_log_debug(D_PLUGINSD, "requested a " PLUGINSD_KEYWORD_FLUSH);
- pluginsd_clear_scope_chart(parser, PLUGINSD_KEYWORD_FLUSH);
- parser->user.replay.start_time = 0;
- parser->user.replay.end_time = 0;
- parser->user.replay.start_time_ut = 0;
- parser->user.replay.end_time_ut = 0;
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_disable(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser) {
- netdata_log_info("PLUGINSD: plugin called DISABLE. Disabling it.");
- parser->user.enabled = 0;
- return PARSER_RC_STOP;
-}
-
-static inline PARSER_RC pluginsd_label(char **words, size_t num_words, PARSER *parser) {
- const char *name = get_word(words, num_words, 1);
- const char *label_source = get_word(words, num_words, 2);
- const char *value = get_word(words, num_words, 3);
-
- if (!name || !label_source || !value)
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_LABEL, "missing parameters");
-
- char *store = (char *)value;
- bool allocated_store = false;
-
- if(unlikely(num_words > 4)) {
- allocated_store = true;
- store = mallocz(PLUGINSD_LINE_MAX + 1);
- size_t remaining = PLUGINSD_LINE_MAX;
- char *move = store;
- char *word;
- for(size_t i = 3; i < num_words && remaining > 2 && (word = get_word(words, num_words, i)) ;i++) {
- if(i > 3) {
- *move++ = ' ';
- *move = '\0';
- remaining--;
- }
-
- size_t length = strlen(word);
- if (length > remaining)
- length = remaining;
-
- remaining -= length;
- memcpy(move, word, length);
- move += length;
- *move = '\0';
- }
- }
-
- if(unlikely(!(parser->user.new_host_labels)))
- parser->user.new_host_labels = rrdlabels_create();
-
- if (strcmp(name,HOST_LABEL_IS_EPHEMERAL) == 0) {
- int is_ephemeral = appconfig_test_boolean_value((char *) value);
- if (is_ephemeral) {
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_LABEL);
- if (likely(host))
- rrdhost_option_set(host, RRDHOST_OPTION_EPHEMERAL_HOST);
- }
- }
-
- rrdlabels_add(parser->user.new_host_labels, name, store, str2l(label_source));
-
- if (allocated_store)
- freez(store);
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_overwrite(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser) {
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_OVERWRITE);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- netdata_log_debug(D_PLUGINSD, "requested to OVERWRITE host labels");
-
- if(unlikely(!host->rrdlabels))
- host->rrdlabels = rrdlabels_create();
-
- rrdlabels_migrate_to_these(host->rrdlabels, parser->user.new_host_labels);
- if (rrdhost_option_check(host, RRDHOST_OPTION_EPHEMERAL_HOST))
- rrdlabels_add(host->rrdlabels, HOST_LABEL_IS_EPHEMERAL, "true", RRDLABEL_SRC_CONFIG);
- rrdhost_flag_set(host, RRDHOST_FLAG_METADATA_LABELS | RRDHOST_FLAG_METADATA_UPDATE);
-
- rrdlabels_destroy(parser->user.new_host_labels);
- parser->user.new_host_labels = NULL;
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_clabel(char **words, size_t num_words, PARSER *parser) {
- const char *name = get_word(words, num_words, 1);
- const char *value = get_word(words, num_words, 2);
- const char *label_source = get_word(words, num_words, 3);
-
- if (!name || !value || !label_source) {
- netdata_log_error("Ignoring malformed or empty CHART LABEL command.");
- return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
- }
-
- if(unlikely(!parser->user.chart_rrdlabels_linked_temporarily)) {
- RRDSET *st = pluginsd_get_scope_chart(parser);
- parser->user.chart_rrdlabels_linked_temporarily = st->rrdlabels;
- rrdlabels_unmark_all(parser->user.chart_rrdlabels_linked_temporarily);
- }
-
- rrdlabels_add(parser->user.chart_rrdlabels_linked_temporarily, name, value, str2l(label_source));
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_clabel_commit(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser) {
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_CLABEL_COMMIT);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_CLABEL_COMMIT, PLUGINSD_KEYWORD_BEGIN);
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- netdata_log_debug(D_PLUGINSD, "requested to commit chart labels");
-
- if(!parser->user.chart_rrdlabels_linked_temporarily) {
- netdata_log_error("PLUGINSD: 'host:%s' got CLABEL_COMMIT, without a CHART or BEGIN. Ignoring it.", rrdhost_hostname(host));
- return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
- }
-
- rrdlabels_remove_all_unmarked(parser->user.chart_rrdlabels_linked_temporarily);
-
- rrdset_flag_set(st, RRDSET_FLAG_METADATA_UPDATE);
- rrdhost_flag_set(st->rrdhost, RRDHOST_FLAG_METADATA_UPDATE);
- rrdset_metadata_updated(st);
-
- parser->user.chart_rrdlabels_linked_temporarily = NULL;
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_replay_begin(char **words, size_t num_words, PARSER *parser) {
- int idx = 1;
- ssize_t slot = pluginsd_parse_rrd_slot(words, num_words);
- if(slot >= 0) idx++;
-
- char *id = get_word(words, num_words, idx++);
- char *start_time_str = get_word(words, num_words, idx++);
- char *end_time_str = get_word(words, num_words, idx++);
- char *child_now_str = get_word(words, num_words, idx++);
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_REPLAY_BEGIN);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st;
- if (likely(!id || !*id))
- st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_REPLAY_BEGIN, PLUGINSD_KEYWORD_REPLAY_BEGIN);
- else
- st = pluginsd_rrdset_cache_get_from_slot(parser, host, id, slot, PLUGINSD_KEYWORD_REPLAY_BEGIN);
-
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- if(!pluginsd_set_scope_chart(parser, st, PLUGINSD_KEYWORD_REPLAY_BEGIN))
- return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- if(start_time_str && end_time_str) {
- time_t start_time = (time_t) str2ull_encoded(start_time_str);
- time_t end_time = (time_t) str2ull_encoded(end_time_str);
-
- time_t wall_clock_time = 0, tolerance;
- bool wall_clock_comes_from_child; (void)wall_clock_comes_from_child;
- if(child_now_str) {
- wall_clock_time = (time_t) str2ull_encoded(child_now_str);
- tolerance = st->update_every + 1;
- wall_clock_comes_from_child = true;
- }
-
- if(wall_clock_time <= 0) {
- wall_clock_time = now_realtime_sec();
- tolerance = st->update_every + 5;
- wall_clock_comes_from_child = false;
- }
-
-#ifdef NETDATA_LOG_REPLICATION_REQUESTS
- internal_error(
- (!st->replay.start_streaming && (end_time < st->replay.after || start_time > st->replay.before)),
- "REPLAY ERROR: 'host:%s/chart:%s' got a " PLUGINSD_KEYWORD_REPLAY_BEGIN " from %ld to %ld, which does not match our request (%ld to %ld).",
- rrdhost_hostname(st->rrdhost), rrdset_id(st), start_time, end_time, st->replay.after, st->replay.before);
-
- internal_error(
- true,
- "REPLAY: 'host:%s/chart:%s' got a " PLUGINSD_KEYWORD_REPLAY_BEGIN " from %ld to %ld, child wall clock is %ld (%s), had requested %ld to %ld",
- rrdhost_hostname(st->rrdhost), rrdset_id(st),
- start_time, end_time, wall_clock_time, wall_clock_comes_from_child ? "from child" : "parent time",
- st->replay.after, st->replay.before);
-#endif
-
- if(start_time && end_time && start_time < wall_clock_time + tolerance && end_time < wall_clock_time + tolerance && start_time < end_time) {
- if (unlikely(end_time - start_time != st->update_every))
- rrdset_set_update_every_s(st, end_time - start_time);
-
- st->last_collected_time.tv_sec = end_time;
- st->last_collected_time.tv_usec = 0;
-
- st->last_updated.tv_sec = end_time;
- st->last_updated.tv_usec = 0;
-
- st->counter++;
- st->counter_done++;
-
- // these are only needed for db mode RAM, SAVE, MAP, ALLOC
- st->db.current_entry++;
- if(st->db.current_entry >= st->db.entries)
- st->db.current_entry -= st->db.entries;
-
- parser->user.replay.start_time = start_time;
- parser->user.replay.end_time = end_time;
- parser->user.replay.start_time_ut = (usec_t) start_time * USEC_PER_SEC;
- parser->user.replay.end_time_ut = (usec_t) end_time * USEC_PER_SEC;
- parser->user.replay.wall_clock_time = wall_clock_time;
- parser->user.replay.rset_enabled = true;
-
- return PARSER_RC_OK;
- }
-
- netdata_log_error("PLUGINSD REPLAY ERROR: 'host:%s/chart:%s' got a " PLUGINSD_KEYWORD_REPLAY_BEGIN
- " from %ld to %ld, but timestamps are invalid "
- "(now is %ld [%s], tolerance %ld). Ignoring " PLUGINSD_KEYWORD_REPLAY_SET,
- rrdhost_hostname(st->rrdhost), rrdset_id(st), start_time, end_time,
- wall_clock_time, wall_clock_comes_from_child ? "child wall clock" : "parent wall clock",
- tolerance);
- }
-
- // the child sends an RBEGIN without any parameters initially
- // setting rset_enabled to false, means the RSET should not store any metrics
- // to store metrics, the RBEGIN needs to have timestamps
- parser->user.replay.start_time = 0;
- parser->user.replay.end_time = 0;
- parser->user.replay.start_time_ut = 0;
- parser->user.replay.end_time_ut = 0;
- parser->user.replay.wall_clock_time = 0;
- parser->user.replay.rset_enabled = false;
- return PARSER_RC_OK;
-}
-
-static inline SN_FLAGS pluginsd_parse_storage_number_flags(const char *flags_str) {
- SN_FLAGS flags = SN_FLAG_NONE;
-
- char c;
- while ((c = *flags_str++)) {
- switch (c) {
- case 'A':
- flags |= SN_FLAG_NOT_ANOMALOUS;
- break;
-
- case 'R':
- flags |= SN_FLAG_RESET;
- break;
-
- case 'E':
- flags = SN_EMPTY_SLOT;
- return flags;
-
- default:
- internal_error(true, "Unknown SN_FLAGS flag '%c'", c);
- break;
- }
- }
-
- return flags;
-}
-
-static inline PARSER_RC pluginsd_replay_set(char **words, size_t num_words, PARSER *parser) {
- int idx = 1;
- ssize_t slot = pluginsd_parse_rrd_slot(words, num_words);
- if(slot >= 0) idx++;
-
- char *dimension = get_word(words, num_words, idx++);
- char *value_str = get_word(words, num_words, idx++);
- char *flags_str = get_word(words, num_words, idx++);
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_REPLAY_SET);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_REPLAY_SET, PLUGINSD_KEYWORD_REPLAY_BEGIN);
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- if(!parser->user.replay.rset_enabled) {
- nd_log_limit_static_thread_var(erl, 1, 0);
- nd_log_limit(&erl, NDLS_COLLECTORS, NDLP_ERR,
- "PLUGINSD: 'host:%s/chart:%s' got a %s but it is disabled by %s errors",
- rrdhost_hostname(host), rrdset_id(st), PLUGINSD_KEYWORD_REPLAY_SET, PLUGINSD_KEYWORD_REPLAY_BEGIN);
-
- // we have to return OK here
- return PARSER_RC_OK;
- }
-
- RRDDIM *rd = pluginsd_acquire_dimension(host, st, dimension, slot, PLUGINSD_KEYWORD_REPLAY_SET);
- if(!rd) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- st->pluginsd.set = true;
-
- if (unlikely(!parser->user.replay.start_time || !parser->user.replay.end_time)) {
- netdata_log_error("PLUGINSD: 'host:%s/chart:%s/dim:%s' got a %s with invalid timestamps %ld to %ld from a %s. Disabling it.",
- rrdhost_hostname(host),
- rrdset_id(st),
- dimension,
- PLUGINSD_KEYWORD_REPLAY_SET,
- parser->user.replay.start_time,
- parser->user.replay.end_time,
- PLUGINSD_KEYWORD_REPLAY_BEGIN);
- return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
- }
-
- if (unlikely(!value_str || !*value_str))
- value_str = "NAN";
-
- if(unlikely(!flags_str))
- flags_str = "";
-
- if (likely(value_str)) {
- RRDDIM_FLAGS rd_flags = rrddim_flag_check(rd, RRDDIM_FLAG_OBSOLETE | RRDDIM_FLAG_ARCHIVED);
-
- if(!(rd_flags & RRDDIM_FLAG_ARCHIVED)) {
- NETDATA_DOUBLE value = str2ndd_encoded(value_str, NULL);
- SN_FLAGS flags = pluginsd_parse_storage_number_flags(flags_str);
-
- if (!netdata_double_isnumber(value) || (flags == SN_EMPTY_SLOT)) {
- value = NAN;
- flags = SN_EMPTY_SLOT;
- }
-
- rrddim_store_metric(rd, parser->user.replay.end_time_ut, value, flags);
- rd->collector.last_collected_time.tv_sec = parser->user.replay.end_time;
- rd->collector.last_collected_time.tv_usec = 0;
- rd->collector.counter++;
- }
- else {
- nd_log_limit_static_global_var(erl, 1, 0);
- nd_log_limit(&erl, NDLS_COLLECTORS, NDLP_WARNING,
- "PLUGINSD: 'host:%s/chart:%s/dim:%s' has the ARCHIVED flag set, but it is replicated. "
- "Ignoring data.",
- rrdhost_hostname(st->rrdhost), rrdset_id(st), rrddim_name(rd));
- }
- }
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_replay_rrddim_collection_state(char **words, size_t num_words, PARSER *parser) {
- if(parser->user.replay.rset_enabled == false)
- return PARSER_RC_OK;
-
- int idx = 1;
- ssize_t slot = pluginsd_parse_rrd_slot(words, num_words);
- if(slot >= 0) idx++;
-
- char *dimension = get_word(words, num_words, idx++);
- char *last_collected_ut_str = get_word(words, num_words, idx++);
- char *last_collected_value_str = get_word(words, num_words, idx++);
- char *last_calculated_value_str = get_word(words, num_words, idx++);
- char *last_stored_value_str = get_word(words, num_words, idx++);
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_REPLAY_RRDDIM_STATE);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_REPLAY_RRDDIM_STATE, PLUGINSD_KEYWORD_REPLAY_BEGIN);
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- if(st->pluginsd.set) {
- // reset pos to reuse the same RDAs
- st->pluginsd.pos = 0;
- st->pluginsd.set = false;
- }
-
- RRDDIM *rd = pluginsd_acquire_dimension(host, st, dimension, slot, PLUGINSD_KEYWORD_REPLAY_RRDDIM_STATE);
- if(!rd) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- usec_t dim_last_collected_ut = (usec_t)rd->collector.last_collected_time.tv_sec * USEC_PER_SEC + (usec_t)rd->collector.last_collected_time.tv_usec;
- usec_t last_collected_ut = last_collected_ut_str ? str2ull_encoded(last_collected_ut_str) : 0;
- if(last_collected_ut > dim_last_collected_ut) {
- rd->collector.last_collected_time.tv_sec = (time_t)(last_collected_ut / USEC_PER_SEC);
- rd->collector.last_collected_time.tv_usec = (last_collected_ut % USEC_PER_SEC);
- }
-
- rd->collector.last_collected_value = last_collected_value_str ? str2ll_encoded(last_collected_value_str) : 0;
- rd->collector.last_calculated_value = last_calculated_value_str ? str2ndd_encoded(last_calculated_value_str, NULL) : 0;
- rd->collector.last_stored_value = last_stored_value_str ? str2ndd_encoded(last_stored_value_str, NULL) : 0.0;
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_replay_rrdset_collection_state(char **words, size_t num_words, PARSER *parser) {
- if(parser->user.replay.rset_enabled == false)
- return PARSER_RC_OK;
-
- char *last_collected_ut_str = get_word(words, num_words, 1);
- char *last_updated_ut_str = get_word(words, num_words, 2);
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_REPLAY_RRDSET_STATE);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_REPLAY_RRDSET_STATE,
- PLUGINSD_KEYWORD_REPLAY_BEGIN);
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- usec_t chart_last_collected_ut = (usec_t)st->last_collected_time.tv_sec * USEC_PER_SEC + (usec_t)st->last_collected_time.tv_usec;
- usec_t last_collected_ut = last_collected_ut_str ? str2ull_encoded(last_collected_ut_str) : 0;
- if(last_collected_ut > chart_last_collected_ut) {
- st->last_collected_time.tv_sec = (time_t)(last_collected_ut / USEC_PER_SEC);
- st->last_collected_time.tv_usec = (last_collected_ut % USEC_PER_SEC);
- }
-
- usec_t chart_last_updated_ut = (usec_t)st->last_updated.tv_sec * USEC_PER_SEC + (usec_t)st->last_updated.tv_usec;
- usec_t last_updated_ut = last_updated_ut_str ? str2ull_encoded(last_updated_ut_str) : 0;
- if(last_updated_ut > chart_last_updated_ut) {
- st->last_updated.tv_sec = (time_t)(last_updated_ut / USEC_PER_SEC);
- st->last_updated.tv_usec = (last_updated_ut % USEC_PER_SEC);
- }
-
- st->counter++;
- st->counter_done++;
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_replay_end(char **words, size_t num_words, PARSER *parser) {
- if (num_words < 7) { // accepts 7, but the 7th is optional
- netdata_log_error("REPLAY: malformed " PLUGINSD_KEYWORD_REPLAY_END " command");
- return PARSER_RC_ERROR;
- }
-
- const char *update_every_child_txt = get_word(words, num_words, 1);
- const char *first_entry_child_txt = get_word(words, num_words, 2);
- const char *last_entry_child_txt = get_word(words, num_words, 3);
- const char *start_streaming_txt = get_word(words, num_words, 4);
- const char *first_entry_requested_txt = get_word(words, num_words, 5);
- const char *last_entry_requested_txt = get_word(words, num_words, 6);
- const char *child_world_time_txt = get_word(words, num_words, 7); // optional
-
- time_t update_every_child = (time_t) str2ull_encoded(update_every_child_txt);
- time_t first_entry_child = (time_t) str2ull_encoded(first_entry_child_txt);
- time_t last_entry_child = (time_t) str2ull_encoded(last_entry_child_txt);
-
- bool start_streaming = (strcmp(start_streaming_txt, "true") == 0);
- time_t first_entry_requested = (time_t) str2ull_encoded(first_entry_requested_txt);
- time_t last_entry_requested = (time_t) str2ull_encoded(last_entry_requested_txt);
-
- // the optional child world time
- time_t child_world_time = (child_world_time_txt && *child_world_time_txt) ? (time_t) str2ull_encoded(
- child_world_time_txt) : now_realtime_sec();
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_REPLAY_END);
- if(!host) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_REPLAY_END, PLUGINSD_KEYWORD_REPLAY_BEGIN);
- if(!st) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
-#ifdef NETDATA_LOG_REPLICATION_REQUESTS
- internal_error(true,
- "PLUGINSD REPLAY: 'host:%s/chart:%s': got a " PLUGINSD_KEYWORD_REPLAY_END " child db from %llu to %llu, start_streaming %s, had requested from %llu to %llu, wall clock %llu",
- rrdhost_hostname(host), rrdset_id(st),
- (unsigned long long)first_entry_child, (unsigned long long)last_entry_child,
- start_streaming?"true":"false",
- (unsigned long long)first_entry_requested, (unsigned long long)last_entry_requested,
- (unsigned long long)child_world_time
- );
-#endif
-
- parser->user.data_collections_count++;
-
- if(parser->user.replay.rset_enabled && st->rrdhost->receiver) {
- time_t now = now_realtime_sec();
- time_t started = st->rrdhost->receiver->replication_first_time_t;
- time_t current = parser->user.replay.end_time;
-
- if(started && current > started) {
- host->rrdpush_receiver_replication_percent = (NETDATA_DOUBLE) (current - started) * 100.0 / (NETDATA_DOUBLE) (now - started);
- worker_set_metric(WORKER_RECEIVER_JOB_REPLICATION_COMPLETION,
- host->rrdpush_receiver_replication_percent);
- }
- }
-
- parser->user.replay.start_time = 0;
- parser->user.replay.end_time = 0;
- parser->user.replay.start_time_ut = 0;
- parser->user.replay.end_time_ut = 0;
- parser->user.replay.wall_clock_time = 0;
- parser->user.replay.rset_enabled = false;
-
- st->counter++;
- st->counter_done++;
- store_metric_collection_completed();
-
-#ifdef NETDATA_LOG_REPLICATION_REQUESTS
- st->replay.start_streaming = false;
- st->replay.after = 0;
- st->replay.before = 0;
- if(start_streaming)
- st->replay.log_next_data_collection = true;
-#endif
-
- if (start_streaming) {
- if (st->update_every != update_every_child)
- rrdset_set_update_every_s(st, update_every_child);
-
- if(rrdset_flag_check(st, RRDSET_FLAG_RECEIVER_REPLICATION_IN_PROGRESS)) {
- rrdset_flag_set(st, RRDSET_FLAG_RECEIVER_REPLICATION_FINISHED);
- rrdset_flag_clear(st, RRDSET_FLAG_RECEIVER_REPLICATION_IN_PROGRESS);
- rrdset_flag_clear(st, RRDSET_FLAG_SYNC_CLOCK);
- rrdhost_receiver_replicating_charts_minus_one(st->rrdhost);
- }
-#ifdef NETDATA_LOG_REPLICATION_REQUESTS
- else
- internal_error(true, "REPLAY ERROR: 'host:%s/chart:%s' got a " PLUGINSD_KEYWORD_REPLAY_END " with enable_streaming = true, but there is no replication in progress for this chart.",
- rrdhost_hostname(host), rrdset_id(st));
-#endif
-
- pluginsd_clear_scope_chart(parser, PLUGINSD_KEYWORD_REPLAY_END);
-
- host->rrdpush_receiver_replication_percent = 100.0;
- worker_set_metric(WORKER_RECEIVER_JOB_REPLICATION_COMPLETION, host->rrdpush_receiver_replication_percent);
-
- return PARSER_RC_OK;
- }
-
- pluginsd_clear_scope_chart(parser, PLUGINSD_KEYWORD_REPLAY_END);
-
- rrdcontext_updated_retention_rrdset(st);
-
- bool ok = replicate_chart_request(send_to_plugin, parser, host, st,
- first_entry_child, last_entry_child, child_world_time,
- first_entry_requested, last_entry_requested);
- return ok ? PARSER_RC_OK : PARSER_RC_ERROR;
-}
-
-static inline PARSER_RC pluginsd_begin_v2(char **words, size_t num_words, PARSER *parser) {
- timing_init();
-
- int idx = 1;
- ssize_t slot = pluginsd_parse_rrd_slot(words, num_words);
- if(slot >= 0) idx++;
-
- char *id = get_word(words, num_words, idx++);
- char *update_every_str = get_word(words, num_words, idx++);
- char *end_time_str = get_word(words, num_words, idx++);
- char *wall_clock_time_str = get_word(words, num_words, idx++);
-
- if(unlikely(!id || !update_every_str || !end_time_str || !wall_clock_time_str))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_BEGIN_V2, "missing parameters");
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_BEGIN_V2);
- if(unlikely(!host)) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- timing_step(TIMING_STEP_BEGIN2_PREPARE);
-
- RRDSET *st = pluginsd_rrdset_cache_get_from_slot(parser, host, id, slot, PLUGINSD_KEYWORD_BEGIN_V2);
-
- if(unlikely(!st)) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- if(!pluginsd_set_scope_chart(parser, st, PLUGINSD_KEYWORD_BEGIN_V2))
- return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- if(unlikely(rrdset_flag_check(st, RRDSET_FLAG_OBSOLETE)))
- rrdset_isnot_obsolete___safe_from_collector_thread(st);
-
- timing_step(TIMING_STEP_BEGIN2_FIND_CHART);
-
- // ------------------------------------------------------------------------
- // parse the parameters
-
- time_t update_every = (time_t) str2ull_encoded(update_every_str);
- time_t end_time = (time_t) str2ull_encoded(end_time_str);
-
- time_t wall_clock_time;
- if(likely(*wall_clock_time_str == '#'))
- wall_clock_time = end_time;
- else
- wall_clock_time = (time_t) str2ull_encoded(wall_clock_time_str);
-
- if (unlikely(update_every != st->update_every))
- rrdset_set_update_every_s(st, update_every);
-
- timing_step(TIMING_STEP_BEGIN2_PARSE);
-
- // ------------------------------------------------------------------------
- // prepare our state
-
- pluginsd_lock_rrdset_data_collection(parser);
-
- parser->user.v2.update_every = update_every;
- parser->user.v2.end_time = end_time;
- parser->user.v2.wall_clock_time = wall_clock_time;
- parser->user.v2.ml_locked = ml_chart_update_begin(st);
-
- timing_step(TIMING_STEP_BEGIN2_ML);
-
- // ------------------------------------------------------------------------
- // propagate it forward in v2
-
- if(!parser->user.v2.stream_buffer.wb && rrdhost_has_rrdpush_sender_enabled(st->rrdhost))
- parser->user.v2.stream_buffer = rrdset_push_metric_initialize(parser->user.st, wall_clock_time);
-
- if(parser->user.v2.stream_buffer.v2 && parser->user.v2.stream_buffer.wb) {
- // check receiver capabilities
- bool can_copy = stream_has_capability(&parser->user, STREAM_CAP_IEEE754) == stream_has_capability(&parser->user.v2.stream_buffer, STREAM_CAP_IEEE754);
-
- // check sender capabilities
- bool with_slots = stream_has_capability(&parser->user.v2.stream_buffer, STREAM_CAP_SLOTS) ? true : false;
- NUMBER_ENCODING integer_encoding = stream_has_capability(&parser->user.v2.stream_buffer, STREAM_CAP_IEEE754) ? NUMBER_ENCODING_BASE64 : NUMBER_ENCODING_HEX;
-
- BUFFER *wb = parser->user.v2.stream_buffer.wb;
-
- buffer_need_bytes(wb, 1024);
-
- if(unlikely(parser->user.v2.stream_buffer.begin_v2_added))
- buffer_fast_strcat(wb, PLUGINSD_KEYWORD_END_V2 "\n", sizeof(PLUGINSD_KEYWORD_END_V2) - 1 + 1);
-
- buffer_fast_strcat(wb, PLUGINSD_KEYWORD_BEGIN_V2, sizeof(PLUGINSD_KEYWORD_BEGIN_V2) - 1);
-
- if(with_slots) {
- buffer_fast_strcat(wb, " "PLUGINSD_KEYWORD_SLOT":", sizeof(PLUGINSD_KEYWORD_SLOT) - 1 + 2);
- buffer_print_uint64_encoded(wb, integer_encoding, st->rrdpush.sender.chart_slot);
- }
-
- buffer_fast_strcat(wb, " '", 2);
- buffer_fast_strcat(wb, rrdset_id(st), string_strlen(st->id));
- buffer_fast_strcat(wb, "' ", 2);
-
- if(can_copy)
- buffer_strcat(wb, update_every_str);
- else
- buffer_print_uint64_encoded(wb, integer_encoding, update_every);
-
- buffer_fast_strcat(wb, " ", 1);
-
- if(can_copy)
- buffer_strcat(wb, end_time_str);
- else
- buffer_print_uint64_encoded(wb, integer_encoding, end_time);
-
- buffer_fast_strcat(wb, " ", 1);
-
- if(can_copy)
- buffer_strcat(wb, wall_clock_time_str);
- else
- buffer_print_uint64_encoded(wb, integer_encoding, wall_clock_time);
-
- buffer_fast_strcat(wb, "\n", 1);
-
- parser->user.v2.stream_buffer.last_point_end_time_s = end_time;
- parser->user.v2.stream_buffer.begin_v2_added = true;
- }
-
- timing_step(TIMING_STEP_BEGIN2_PROPAGATE);
-
- // ------------------------------------------------------------------------
- // store it
-
- st->last_collected_time.tv_sec = end_time;
- st->last_collected_time.tv_usec = 0;
- st->last_updated.tv_sec = end_time;
- st->last_updated.tv_usec = 0;
- st->counter++;
- st->counter_done++;
-
- // these are only needed for db mode RAM, SAVE, MAP, ALLOC
- st->db.current_entry++;
- if(st->db.current_entry >= st->db.entries)
- st->db.current_entry -= st->db.entries;
-
- timing_step(TIMING_STEP_BEGIN2_STORE);
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_set_v2(char **words, size_t num_words, PARSER *parser) {
- timing_init();
-
- int idx = 1;
- ssize_t slot = pluginsd_parse_rrd_slot(words, num_words);
- if(slot >= 0) idx++;
-
- char *dimension = get_word(words, num_words, idx++);
- char *collected_str = get_word(words, num_words, idx++);
- char *value_str = get_word(words, num_words, idx++);
- char *flags_str = get_word(words, num_words, idx++);
-
- if(unlikely(!dimension || !collected_str || !value_str || !flags_str))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_SET_V2, "missing parameters");
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_SET_V2);
- if(unlikely(!host)) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_SET_V2, PLUGINSD_KEYWORD_BEGIN_V2);
- if(unlikely(!st)) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- timing_step(TIMING_STEP_SET2_PREPARE);
-
- RRDDIM *rd = pluginsd_acquire_dimension(host, st, dimension, slot, PLUGINSD_KEYWORD_SET_V2);
- if(unlikely(!rd)) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- st->pluginsd.set = true;
-
- if(unlikely(rrddim_flag_check(rd, RRDDIM_FLAG_OBSOLETE | RRDDIM_FLAG_ARCHIVED)))
- rrddim_isnot_obsolete___safe_from_collector_thread(st, rd);
-
- timing_step(TIMING_STEP_SET2_LOOKUP_DIMENSION);
-
- // ------------------------------------------------------------------------
- // parse the parameters
-
- collected_number collected_value = (collected_number) str2ll_encoded(collected_str);
-
- NETDATA_DOUBLE value;
- if(*value_str == '#')
- value = (NETDATA_DOUBLE)collected_value;
- else
- value = str2ndd_encoded(value_str, NULL);
-
- SN_FLAGS flags = pluginsd_parse_storage_number_flags(flags_str);
-
- timing_step(TIMING_STEP_SET2_PARSE);
-
- // ------------------------------------------------------------------------
- // check value and ML
-
- if (unlikely(!netdata_double_isnumber(value) || (flags == SN_EMPTY_SLOT))) {
- value = NAN;
- flags = SN_EMPTY_SLOT;
-
- if(parser->user.v2.ml_locked)
- ml_dimension_is_anomalous(rd, parser->user.v2.end_time, 0, false);
- }
- else if(parser->user.v2.ml_locked) {
- if (ml_dimension_is_anomalous(rd, parser->user.v2.end_time, value, true)) {
- // clear anomaly bit: 0 -> is anomalous, 1 -> not anomalous
- flags &= ~((storage_number) SN_FLAG_NOT_ANOMALOUS);
- }
- else
- flags |= SN_FLAG_NOT_ANOMALOUS;
- }
-
- timing_step(TIMING_STEP_SET2_ML);
-
- // ------------------------------------------------------------------------
- // propagate it forward in v2
-
- if(parser->user.v2.stream_buffer.v2 && parser->user.v2.stream_buffer.begin_v2_added && parser->user.v2.stream_buffer.wb) {
- // check if receiver and sender have the same number parsing capabilities
- bool can_copy = stream_has_capability(&parser->user, STREAM_CAP_IEEE754) == stream_has_capability(&parser->user.v2.stream_buffer, STREAM_CAP_IEEE754);
-
- // check the sender capabilities
- bool with_slots = stream_has_capability(&parser->user.v2.stream_buffer, STREAM_CAP_SLOTS) ? true : false;
- NUMBER_ENCODING integer_encoding = stream_has_capability(&parser->user.v2.stream_buffer, STREAM_CAP_IEEE754) ? NUMBER_ENCODING_BASE64 : NUMBER_ENCODING_HEX;
- NUMBER_ENCODING doubles_encoding = stream_has_capability(&parser->user.v2.stream_buffer, STREAM_CAP_IEEE754) ? NUMBER_ENCODING_BASE64 : NUMBER_ENCODING_DECIMAL;
-
- BUFFER *wb = parser->user.v2.stream_buffer.wb;
- buffer_need_bytes(wb, 1024);
- buffer_fast_strcat(wb, PLUGINSD_KEYWORD_SET_V2, sizeof(PLUGINSD_KEYWORD_SET_V2) - 1);
-
- if(with_slots) {
- buffer_fast_strcat(wb, " "PLUGINSD_KEYWORD_SLOT":", sizeof(PLUGINSD_KEYWORD_SLOT) - 1 + 2);
- buffer_print_uint64_encoded(wb, integer_encoding, rd->rrdpush.sender.dim_slot);
- }
-
- buffer_fast_strcat(wb, " '", 2);
- buffer_fast_strcat(wb, rrddim_id(rd), string_strlen(rd->id));
- buffer_fast_strcat(wb, "' ", 2);
- if(can_copy)
- buffer_strcat(wb, collected_str);
- else
- buffer_print_int64_encoded(wb, integer_encoding, collected_value); // original v2 had hex
- buffer_fast_strcat(wb, " ", 1);
- if(can_copy)
- buffer_strcat(wb, value_str);
- else
- buffer_print_netdata_double_encoded(wb, doubles_encoding, value); // original v2 had decimal
- buffer_fast_strcat(wb, " ", 1);
- buffer_print_sn_flags(wb, flags, true);
- buffer_fast_strcat(wb, "\n", 1);
- }
-
- timing_step(TIMING_STEP_SET2_PROPAGATE);
-
- // ------------------------------------------------------------------------
- // store it
-
- rrddim_store_metric(rd, parser->user.v2.end_time * USEC_PER_SEC, value, flags);
- rd->collector.last_collected_time.tv_sec = parser->user.v2.end_time;
- rd->collector.last_collected_time.tv_usec = 0;
- rd->collector.last_collected_value = collected_value;
- rd->collector.last_stored_value = value;
- rd->collector.last_calculated_value = value;
- rd->collector.counter++;
- rrddim_set_updated(rd);
-
- timing_step(TIMING_STEP_SET2_STORE);
-
- return PARSER_RC_OK;
-}
-
-void pluginsd_cleanup_v2(PARSER *parser) {
- // this is called when the thread is stopped while processing
- pluginsd_clear_scope_chart(parser, "THREAD CLEANUP");
-}
-
-static inline PARSER_RC pluginsd_end_v2(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser) {
- timing_init();
-
- RRDHOST *host = pluginsd_require_scope_host(parser, PLUGINSD_KEYWORD_END_V2);
- if(unlikely(!host)) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- RRDSET *st = pluginsd_require_scope_chart(parser, PLUGINSD_KEYWORD_END_V2, PLUGINSD_KEYWORD_BEGIN_V2);
- if(unlikely(!st)) return PLUGINSD_DISABLE_PLUGIN(parser, NULL, NULL);
-
- parser->user.data_collections_count++;
-
- timing_step(TIMING_STEP_END2_PREPARE);
-
- // ------------------------------------------------------------------------
- // propagate the whole chart update in v1
-
- if(unlikely(!parser->user.v2.stream_buffer.v2 && !parser->user.v2.stream_buffer.begin_v2_added && parser->user.v2.stream_buffer.wb))
- rrdset_push_metrics_v1(&parser->user.v2.stream_buffer, st);
-
- timing_step(TIMING_STEP_END2_PUSH_V1);
-
- // ------------------------------------------------------------------------
- // unblock data collection
-
- pluginsd_unlock_previous_scope_chart(parser, PLUGINSD_KEYWORD_END_V2, false);
- rrdcontext_collected_rrdset(st);
- store_metric_collection_completed();
-
- timing_step(TIMING_STEP_END2_RRDSET);
-
- // ------------------------------------------------------------------------
- // propagate it forward
-
- rrdset_push_metrics_finished(&parser->user.v2.stream_buffer, st);
-
- timing_step(TIMING_STEP_END2_PROPAGATE);
-
- // ------------------------------------------------------------------------
- // cleanup RRDSET / RRDDIM
-
- if(likely(st->pluginsd.dims_with_slots)) {
- for(size_t i = 0; i < st->pluginsd.size ;i++) {
- RRDDIM *rd = st->pluginsd.prd_array[i].rd;
-
- if(!rd)
- continue;
-
- rd->collector.calculated_value = 0;
- rd->collector.collected_value = 0;
- rrddim_clear_updated(rd);
- }
- }
- else {
- RRDDIM *rd;
- rrddim_foreach_read(rd, st){
- rd->collector.calculated_value = 0;
- rd->collector.collected_value = 0;
- rrddim_clear_updated(rd);
- }
- rrddim_foreach_done(rd);
- }
-
- // ------------------------------------------------------------------------
- // reset state
-
- parser->user.v2 = (struct parser_user_object_v2){ 0 };
-
- timing_step(TIMING_STEP_END2_STORE);
- timing_report();
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_exit(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser __maybe_unused) {
- netdata_log_info("PLUGINSD: plugin called EXIT.");
- return PARSER_RC_STOP;
-}
-
-struct mutex_cond {
- pthread_mutex_t lock;
- pthread_cond_t cond;
- int rc;
-};
-
-static void virt_fnc_got_data_cb(BUFFER *wb __maybe_unused, int code, void *callback_data)
-{
- struct mutex_cond *ctx = callback_data;
- pthread_mutex_lock(&ctx->lock);
- ctx->rc = code;
- pthread_cond_broadcast(&ctx->cond);
- pthread_mutex_unlock(&ctx->lock);
-}
-
-#define VIRT_FNC_TIMEOUT 1
-#define VIRT_FNC_BUF_SIZE (4096)
-void call_virtual_function_async(BUFFER *wb, RRDHOST *host, const char *name, const char *payload, rrd_function_result_callback_t callback, void *callback_data) {
- PARSER *parser = NULL;
-
- //TODO simplify (as we really need only first parameter to get plugin name maybe we can avoid parsing all)
- char *words[PLUGINSD_MAX_WORDS];
- char *function_with_params = strdupz(name);
- size_t num_words = quoted_strings_splitter(function_with_params, words, PLUGINSD_MAX_WORDS, isspace_map_pluginsd);
-
- if (num_words < 2) {
- netdata_log_error("PLUGINSD: virtual function name is empty.");
- freez(function_with_params);
- return;
- }
-
- const DICTIONARY_ITEM *cpi = dictionary_get_and_acquire_item(host->configurable_plugins, get_word(words, num_words, 1));
- if (unlikely(cpi == NULL)) {
- netdata_log_error("PLUGINSD: virtual function plugin '%s' not found.", name);
- freez(function_with_params);
- return;
- }
- struct configurable_plugin *cp = dictionary_acquired_item_value(cpi);
- parser = (PARSER *)cp->cb_usr_ctx;
-
- BUFFER *function_out = buffer_create(VIRT_FNC_BUF_SIZE, NULL);
- // if we are forwarding this to a plugin (as opposed to streaming/child) we have to remove the first parameter (plugin_name)
- buffer_strcat(function_out, get_word(words, num_words, 0));
- for (size_t i = 1; i < num_words; i++) {
- if (i == 1 && SERVING_PLUGINSD(parser))
- continue;
- buffer_sprintf(function_out, " %s", get_word(words, num_words, i));
- }
- freez(function_with_params);
-
- usec_t now = now_realtime_usec();
-
- struct inflight_function tmp = {
- .started_ut = now,
- .timeout_ut = now + VIRT_FNC_TIMEOUT + USEC_PER_SEC,
- .result_body_wb = wb,
- .timeout = VIRT_FNC_TIMEOUT * 10,
- .function = string_strdupz(buffer_tostring(function_out)),
- .result_cb = callback,
- .result_cb_data = callback_data,
- .payload = payload != NULL ? strdupz(payload) : NULL,
- .virtual = true,
- };
- buffer_free(function_out);
-
- uuid_t uuid;
- uuid_generate_time(uuid);
-
- char key[UUID_STR_LEN];
- uuid_unparse_lower(uuid, key);
-
- dictionary_write_lock(parser->inflight.functions);
-
- // if there is any error, our dictionary callbacks will call the caller callback to notify
- // the caller about the error - no need for error handling here.
- dictionary_set(parser->inflight.functions, key, &tmp, sizeof(struct inflight_function));
-
- if(!parser->inflight.smaller_timeout || tmp.timeout_ut < parser->inflight.smaller_timeout)
- parser->inflight.smaller_timeout = tmp.timeout_ut;
-
- // garbage collect stale inflight functions
- if(parser->inflight.smaller_timeout < now)
- inflight_functions_garbage_collect(parser, now);
-
- dictionary_write_unlock(parser->inflight.functions);
-}
-
-
-dyncfg_config_t call_virtual_function_blocking(PARSER *parser, const char *name, int *rc, const char *payload) {
- usec_t now = now_realtime_usec();
- BUFFER *wb = buffer_create(VIRT_FNC_BUF_SIZE, NULL);
-
- struct mutex_cond cond = {
- .lock = PTHREAD_MUTEX_INITIALIZER,
- .cond = PTHREAD_COND_INITIALIZER
- };
-
- struct inflight_function tmp = {
- .started_ut = now,
- .timeout_ut = now + VIRT_FNC_TIMEOUT + USEC_PER_SEC,
- .result_body_wb = wb,
- .timeout = VIRT_FNC_TIMEOUT,
- .function = string_strdupz(name),
- .result_cb = virt_fnc_got_data_cb,
- .result_cb_data = &cond,
- .payload = payload != NULL ? strdupz(payload) : NULL,
- .virtual = true,
- };
-
- uuid_t uuid;
- uuid_generate_time(uuid);
-
- char key[UUID_STR_LEN];
- uuid_unparse_lower(uuid, key);
-
- dictionary_write_lock(parser->inflight.functions);
-
- // if there is any error, our dictionary callbacks will call the caller callback to notify
- // the caller about the error - no need for error handling here.
- dictionary_set(parser->inflight.functions, key, &tmp, sizeof(struct inflight_function));
-
- if(!parser->inflight.smaller_timeout || tmp.timeout_ut < parser->inflight.smaller_timeout)
- parser->inflight.smaller_timeout = tmp.timeout_ut;
-
- // garbage collect stale inflight functions
- if(parser->inflight.smaller_timeout < now)
- inflight_functions_garbage_collect(parser, now);
-
- dictionary_write_unlock(parser->inflight.functions);
-
- struct timespec tp;
- clock_gettime(CLOCK_REALTIME, &tp);
- tp.tv_sec += (time_t)VIRT_FNC_TIMEOUT;
-
- pthread_mutex_lock(&cond.lock);
-
- int ret = pthread_cond_timedwait(&cond.cond, &cond.lock, &tp);
- if (ret == ETIMEDOUT)
- netdata_log_error("PLUGINSD: DYNCFG virtual function %s timed out", name);
-
- pthread_mutex_unlock(&cond.lock);
-
- dyncfg_config_t cfg;
- cfg.data = strdupz(buffer_tostring(wb));
- cfg.data_size = buffer_strlen(wb);
-
- if (rc != NULL)
- *rc = cond.rc;
-
- buffer_free(wb);
- return cfg;
-}
-
-#define CVF_MAX_LEN (1024)
-static dyncfg_config_t get_plugin_config_cb(void *usr_ctx, const char *plugin_name)
-{
- PARSER *parser = usr_ctx;
-
- if (SERVING_STREAMING(parser)) {
- char buf[CVF_MAX_LEN + 1];
- snprintfz(buf, CVF_MAX_LEN, FUNCTION_NAME_GET_PLUGIN_CONFIG " %s", plugin_name);
- return call_virtual_function_blocking(parser, buf, NULL, NULL);
- }
-
- return call_virtual_function_blocking(parser, FUNCTION_NAME_GET_PLUGIN_CONFIG, NULL, NULL);
-}
-
-static dyncfg_config_t get_plugin_config_schema_cb(void *usr_ctx, const char *plugin_name)
-{
- PARSER *parser = usr_ctx;
-
- if (SERVING_STREAMING(parser)) {
- char buf[CVF_MAX_LEN + 1];
- snprintfz(buf, CVF_MAX_LEN, FUNCTION_NAME_GET_PLUGIN_CONFIG_SCHEMA " %s", plugin_name);
- return call_virtual_function_blocking(parser, buf, NULL, NULL);
- }
-
- return call_virtual_function_blocking(parser, "get_plugin_config_schema", NULL, NULL);
-}
-
-static dyncfg_config_t get_module_config_cb(void *usr_ctx, const char *plugin_name, const char *module_name)
-{
- PARSER *parser = usr_ctx;
- BUFFER *wb = buffer_create(CVF_MAX_LEN, NULL);
-
- buffer_strcat(wb, FUNCTION_NAME_GET_MODULE_CONFIG);
- if (SERVING_STREAMING(parser))
- buffer_sprintf(wb, " %s", plugin_name);
-
- buffer_sprintf(wb, " %s", module_name);
-
- dyncfg_config_t ret = call_virtual_function_blocking(parser, buffer_tostring(wb), NULL, NULL);
-
- buffer_free(wb);
-
- return ret;
-}
-
-static dyncfg_config_t get_module_config_schema_cb(void *usr_ctx, const char *plugin_name, const char *module_name)
-{
- PARSER *parser = usr_ctx;
- BUFFER *wb = buffer_create(CVF_MAX_LEN, NULL);
-
- buffer_strcat(wb, FUNCTION_NAME_GET_MODULE_CONFIG_SCHEMA);
- if (SERVING_STREAMING(parser))
- buffer_sprintf(wb, " %s", plugin_name);
-
- buffer_sprintf(wb, " %s", module_name);
-
- dyncfg_config_t ret = call_virtual_function_blocking(parser, buffer_tostring(wb), NULL, NULL);
-
- buffer_free(wb);
-
- return ret;
-}
-
-static dyncfg_config_t get_job_config_schema_cb(void *usr_ctx, const char *plugin_name, const char *module_name)
-{
- PARSER *parser = usr_ctx;
- BUFFER *wb = buffer_create(CVF_MAX_LEN, NULL);
-
- buffer_strcat(wb, FUNCTION_NAME_GET_JOB_CONFIG_SCHEMA);
-
- if (SERVING_STREAMING(parser))
- buffer_sprintf(wb, " %s", plugin_name);
-
- buffer_sprintf(wb, " %s", module_name);
-
- dyncfg_config_t ret = call_virtual_function_blocking(parser, buffer_tostring(wb), NULL, NULL);
-
- buffer_free(wb);
-
- return ret;
-}
-
-static dyncfg_config_t get_job_config_cb(void *usr_ctx, const char *plugin_name, const char *module_name, const char* job_name)
-{
- PARSER *parser = usr_ctx;
- BUFFER *wb = buffer_create(CVF_MAX_LEN, NULL);
-
- buffer_strcat(wb, FUNCTION_NAME_GET_JOB_CONFIG);
-
- if (SERVING_STREAMING(parser))
- buffer_sprintf(wb, " %s", plugin_name);
-
- buffer_sprintf(wb, " %s %s", module_name, job_name);
-
- dyncfg_config_t ret = call_virtual_function_blocking(parser, buffer_tostring(wb), NULL, NULL);
-
- buffer_free(wb);
-
- return ret;
-}
-
-enum set_config_result set_plugin_config_cb(void *usr_ctx, const char *plugin_name, dyncfg_config_t *cfg)
-{
- PARSER *parser = usr_ctx;
- BUFFER *wb = buffer_create(CVF_MAX_LEN, NULL);
-
- buffer_strcat(wb, FUNCTION_NAME_SET_PLUGIN_CONFIG);
-
- if (SERVING_STREAMING(parser))
- buffer_sprintf(wb, " %s", plugin_name);
-
- int rc;
- call_virtual_function_blocking(parser, buffer_tostring(wb), &rc, cfg->data);
-
- buffer_free(wb);
- if(rc != DYNCFG_VFNC_RET_CFG_ACCEPTED)
- return SET_CONFIG_REJECTED;
- return SET_CONFIG_ACCEPTED;
-}
-
-enum set_config_result set_module_config_cb(void *usr_ctx, const char *plugin_name, const char *module_name, dyncfg_config_t *cfg)
-{
- PARSER *parser = usr_ctx;
- BUFFER *wb = buffer_create(CVF_MAX_LEN, NULL);
-
- buffer_strcat(wb, FUNCTION_NAME_SET_MODULE_CONFIG);
-
- if (SERVING_STREAMING(parser))
- buffer_sprintf(wb, " %s", plugin_name);
-
- buffer_sprintf(wb, " %s", module_name);
-
- int rc;
- call_virtual_function_blocking(parser, buffer_tostring(wb), &rc, cfg->data);
-
- buffer_free(wb);
-
- if(rc != DYNCFG_VFNC_RET_CFG_ACCEPTED)
- return SET_CONFIG_REJECTED;
- return SET_CONFIG_ACCEPTED;
-}
-
-enum set_config_result set_job_config_cb(void *usr_ctx, const char *plugin_name, const char *module_name, const char *job_name, dyncfg_config_t *cfg)
-{
- PARSER *parser = usr_ctx;
- BUFFER *wb = buffer_create(CVF_MAX_LEN, NULL);
-
- buffer_strcat(wb, FUNCTION_NAME_SET_JOB_CONFIG);
-
- if (SERVING_STREAMING(parser))
- buffer_sprintf(wb, " %s", plugin_name);
-
- buffer_sprintf(wb, " %s %s", module_name, job_name);
-
- int rc;
- call_virtual_function_blocking(parser, buffer_tostring(wb), &rc, cfg->data);
-
- buffer_free(wb);
-
- if(rc != DYNCFG_VFNC_RET_CFG_ACCEPTED)
- return SET_CONFIG_REJECTED;
- return SET_CONFIG_ACCEPTED;
-}
-
-enum set_config_result delete_job_cb(void *usr_ctx, const char *plugin_name ,const char *module_name, const char *job_name)
-{
- PARSER *parser = usr_ctx;
- BUFFER *wb = buffer_create(CVF_MAX_LEN, NULL);
-
- buffer_strcat(wb, FUNCTION_NAME_DELETE_JOB);
-
- if (SERVING_STREAMING(parser))
- buffer_sprintf(wb, " %s", plugin_name);
-
- buffer_sprintf(wb, " %s %s", module_name, job_name);
-
- int rc;
- call_virtual_function_blocking(parser, buffer_tostring(wb), &rc, NULL);
-
- buffer_free(wb);
-
- if(rc != DYNCFG_VFNC_RET_CFG_ACCEPTED)
- return SET_CONFIG_REJECTED;
- return SET_CONFIG_ACCEPTED;
-}
-
-
-static inline PARSER_RC pluginsd_register_plugin(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser __maybe_unused) {
- netdata_log_info("PLUGINSD: DYNCFG_ENABLE");
-
- if (unlikely (num_words != 2))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_ENABLE, "missing name parameter");
-
- struct configurable_plugin *cfg = callocz(1, sizeof(struct configurable_plugin));
-
- cfg->name = strdupz(words[1]);
- cfg->set_config_cb = set_plugin_config_cb;
- cfg->get_config_cb = get_plugin_config_cb;
- cfg->get_config_schema_cb = get_plugin_config_schema_cb;
- cfg->cb_usr_ctx = parser;
-
- const DICTIONARY_ITEM *di = register_plugin(parser->user.host->configurable_plugins, cfg, SERVING_PLUGINSD(parser));
- if (unlikely(di == NULL)) {
- freez(cfg->name);
- freez(cfg);
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_ENABLE, "error registering plugin");
- }
-
- if (SERVING_PLUGINSD(parser)) {
- // this is optimization for pluginsd to avoid extra dictionary lookup
- // as we know which plugin is comunicating with us
- parser->user.cd->cfg_dict_item = di;
- parser->user.cd->configuration = cfg;
- } else {
- // register_plugin keeps the item acquired, so we need to release it
- dictionary_acquired_item_release(parser->user.host->configurable_plugins, di);
- }
-
- rrdpush_send_dyncfg_enable(parser->user.host, cfg->name);
-
- return PARSER_RC_OK;
-}
-
-#define LOG_MSG_SIZE (1024)
-#define MODULE_NAME_IDX (SERVING_PLUGINSD(parser) ? 1 : 2)
-#define MODULE_TYPE_IDX (SERVING_PLUGINSD(parser) ? 2 : 3)
-static inline PARSER_RC pluginsd_register_module(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser __maybe_unused) {
- netdata_log_info("PLUGINSD: DYNCFG_REG_MODULE");
-
- size_t expected_num_words = SERVING_PLUGINSD(parser) ? 3 : 4;
-
- if (unlikely(num_words != expected_num_words)) {
- char log[LOG_MSG_SIZE + 1];
- snprintfz(log, LOG_MSG_SIZE, "expected %zu (got %zu) parameters: %smodule_name module_type", expected_num_words - 1, num_words - 1, SERVING_PLUGINSD(parser) ? "" : "plugin_name ");
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_REGISTER_MODULE, log);
- }
-
- struct configurable_plugin *plug_cfg;
- const DICTIONARY_ITEM *di = NULL;
- if (SERVING_PLUGINSD(parser)) {
- plug_cfg = parser->user.cd->configuration;
- if (unlikely(plug_cfg == NULL))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_REGISTER_MODULE, "you have to enable dynamic configuration first using " PLUGINSD_KEYWORD_DYNCFG_ENABLE);
- } else {
- di = dictionary_get_and_acquire_item(parser->user.host->configurable_plugins, words[1]);
- if (unlikely(di == NULL))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_REGISTER_MODULE, "plugin not found");
-
- plug_cfg = (struct configurable_plugin *)dictionary_acquired_item_value(di);
- }
-
- struct module *mod = callocz(1, sizeof(struct module));
-
- mod->type = str2_module_type(words[MODULE_TYPE_IDX]);
- if (unlikely(mod->type == MOD_TYPE_UNKNOWN)) {
- freez(mod);
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_REGISTER_MODULE, "unknown module type (allowed: job_array, single)");
- }
-
- mod->name = strdupz(words[MODULE_NAME_IDX]);
-
- mod->set_config_cb = set_module_config_cb;
- mod->get_config_cb = get_module_config_cb;
- mod->get_config_schema_cb = get_module_config_schema_cb;
- mod->config_cb_usr_ctx = parser;
-
- mod->get_job_config_cb = get_job_config_cb;
- mod->get_job_config_schema_cb = get_job_config_schema_cb;
- mod->set_job_config_cb = set_job_config_cb;
- mod->delete_job_cb = delete_job_cb;
- mod->job_config_cb_usr_ctx = parser;
-
- register_module(parser->user.host->configurable_plugins, plug_cfg, mod, SERVING_PLUGINSD(parser));
-
- if (di != NULL)
- dictionary_acquired_item_release(parser->user.host->configurable_plugins, di);
-
- rrdpush_send_dyncfg_reg_module(parser->user.host, plug_cfg->name, mod->name, mod->type);
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_register_job_common(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser __maybe_unused, const char *plugin_name) {
- const char *module_name = words[0];
- const char *job_name = words[1];
- const char *job_type_str = words[2];
- const char *flags_str = words[3];
-
- long f = str2l(flags_str);
-
- if (f < 0)
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_REGISTER_JOB, "invalid flags received");
-
- dyncfg_job_flg_t flags = f;
-
- if (SERVING_PLUGINSD(parser))
- flags |= JOB_FLG_PLUGIN_PUSHED;
- else
- flags |= JOB_FLG_STREAMING_PUSHED;
-
- enum job_type job_type = dyncfg_str2job_type(job_type_str);
- if (job_type == JOB_TYPE_UNKNOWN)
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_REGISTER_JOB, "unknown job type");
-
- if (SERVING_PLUGINSD(parser) && job_type == JOB_TYPE_USER)
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_REGISTER_JOB, "plugins cannot push jobs of type \"user\" (this is allowed only in streaming)");
-
- if (register_job(parser->user.host->configurable_plugins, plugin_name, module_name, job_name, job_type, flags, 0)) // ignore existing is off as this is explicitly called register job
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_REGISTER_JOB, "error registering job");
-
- rrdpush_send_dyncfg_reg_job(parser->user.host, plugin_name, module_name, job_name, job_type, flags);
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_register_job(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser __maybe_unused) {
- size_t expected_num_words = SERVING_PLUGINSD(parser) ? 5 : 6;
-
- if (unlikely(num_words != expected_num_words)) {
- char log[LOG_MSG_SIZE + 1];
- snprintfz(log, LOG_MSG_SIZE, "expected %zu (got %zu) parameters: %smodule_name job_name job_type", expected_num_words - 1, num_words - 1, SERVING_PLUGINSD(parser) ? "" : "plugin_name ");
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_REGISTER_JOB, log);
- }
-
- if (SERVING_PLUGINSD(parser)) {
- return pluginsd_register_job_common(&words[1], num_words - 1, parser, parser->user.cd->configuration->name);
- }
- return pluginsd_register_job_common(&words[2], num_words - 2, parser, words[1]);
-}
-
-static inline PARSER_RC pluginsd_dyncfg_reset(char **words __maybe_unused, size_t num_words __maybe_unused, PARSER *parser __maybe_unused) {
- if (unlikely(num_words != (SERVING_PLUGINSD(parser) ? 1 : 2)))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_RESET, SERVING_PLUGINSD(parser) ? "expected 0 parameters" : "expected 1 parameter: plugin_name");
-
- if (SERVING_PLUGINSD(parser)) {
- unregister_plugin(parser->user.host->configurable_plugins, parser->user.cd->cfg_dict_item);
- rrdpush_send_dyncfg_reset(parser->user.host, parser->user.cd->configuration->name);
- parser->user.cd->configuration = NULL;
- } else {
- const DICTIONARY_ITEM *di = dictionary_get_and_acquire_item(parser->user.host->configurable_plugins, words[1]);
- if (unlikely(di == NULL))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DYNCFG_RESET, "plugin not found");
- unregister_plugin(parser->user.host->configurable_plugins, di);
- rrdpush_send_dyncfg_reset(parser->user.host, words[1]);
- }
-
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC pluginsd_job_status_common(char **words, size_t num_words, PARSER *parser, const char *plugin_name) {
- int state = str2i(words[3]);
-
- enum job_status status = str2job_state(words[2]);
- if (unlikely(SERVING_PLUGINSD(parser) && status == JOB_STATUS_UNKNOWN))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_REPORT_JOB_STATUS, "unknown job status");
-
- char *message = NULL;
- if (num_words == 5 && strlen(words[4]) > 0)
- message = words[4];
-
- const DICTIONARY_ITEM *plugin_item;
- DICTIONARY *job_dict;
- const DICTIONARY_ITEM *job_item = report_job_status_acq_lock(parser->user.host->configurable_plugins, &plugin_item, &job_dict, plugin_name, words[0], words[1], status, state, message);
-
- if (job_item != NULL) {
- struct job *job = dictionary_acquired_item_value(job_item);
- rrdpush_send_job_status_update(parser->user.host, plugin_name, words[0], job);
-
- pthread_mutex_unlock(&job->lock);
- dictionary_acquired_item_release(job_dict, job_item);
- dictionary_acquired_item_release(parser->user.host->configurable_plugins, plugin_item);
- }
-
- return PARSER_RC_OK;
-}
-
-// job_status [plugin_name if streaming] <module_name> <job_name> <status_code> <state> [message]
-static PARSER_RC pluginsd_job_status(char **words, size_t num_words, PARSER *parser) {
- if (SERVING_PLUGINSD(parser)) {
- if (unlikely(num_words != 5 && num_words != 6))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_REPORT_JOB_STATUS, "expected 4 or 5 parameters: module_name, job_name, status_code, state, [optional: message]");
- } else {
- if (unlikely(num_words != 6 && num_words != 7))
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_REPORT_JOB_STATUS, "expected 5 or 6 parameters: plugin_name, module_name, job_name, status_code, state, [optional: message]");
- }
-
- if (SERVING_PLUGINSD(parser)) {
- return pluginsd_job_status_common(&words[1], num_words - 1, parser, parser->user.cd->configuration->name);
- }
- return pluginsd_job_status_common(&words[2], num_words - 2, parser, words[1]);
-}
-
-static PARSER_RC pluginsd_delete_job(char **words, size_t num_words, PARSER *parser) {
- // this can confuse a bit but there is a diference between KEYWORD_DELETE_JOB and actual delete_job function
- // they are of opossite direction
- if (num_words != 4)
- return PLUGINSD_DISABLE_PLUGIN(parser, PLUGINSD_KEYWORD_DELETE_JOB, "expected 2 parameters: plugin_name, module_name, job_name");
-
- const char *plugin_name = get_word(words, num_words, 1);
- const char *module_name = get_word(words, num_words, 2);
- const char *job_name = get_word(words, num_words, 3);
-
- if (SERVING_STREAMING(parser))
- delete_job_pname(parser->user.host->configurable_plugins, plugin_name, module_name, job_name);
-
- // forward to parent if any
- rrdpush_send_job_deleted(parser->user.host, plugin_name, module_name, job_name);
- return PARSER_RC_OK;
-}
-
-static inline PARSER_RC streaming_claimed_id(char **words, size_t num_words, PARSER *parser)
-{
- const char *host_uuid_str = get_word(words, num_words, 1);
- const char *claim_id_str = get_word(words, num_words, 2);
-
- if (!host_uuid_str || !claim_id_str) {
- netdata_log_error("Command CLAIMED_ID came malformed, uuid = '%s', claim_id = '%s'",
- host_uuid_str ? host_uuid_str : "[unset]",
- claim_id_str ? claim_id_str : "[unset]");
- return PARSER_RC_ERROR;
- }
-
- uuid_t uuid;
- RRDHOST *host = parser->user.host;
-
- // We don't need the parsed UUID
- // just do it to check the format
- if(uuid_parse(host_uuid_str, uuid)) {
- netdata_log_error("1st parameter (host GUID) to CLAIMED_ID command is not valid GUID. Received: \"%s\".", host_uuid_str);
- return PARSER_RC_ERROR;
- }
- if(uuid_parse(claim_id_str, uuid) && strcmp(claim_id_str, "NULL") != 0) {
- netdata_log_error("2nd parameter (Claim ID) to CLAIMED_ID command is not valid GUID. Received: \"%s\".", claim_id_str);
- return PARSER_RC_ERROR;
- }
-
- if(strcmp(host_uuid_str, host->machine_guid) != 0) {
- netdata_log_error("Claim ID is for host \"%s\" but it came over connection for \"%s\"", host_uuid_str, host->machine_guid);
- return PARSER_RC_OK; //the message is OK problem must be somewhere else
- }
-
- rrdhost_aclk_state_lock(host);
-
- if (host->aclk_state.claimed_id)
- freez(host->aclk_state.claimed_id);
-
- host->aclk_state.claimed_id = strcmp(claim_id_str, "NULL") ? strdupz(claim_id_str) : NULL;
-
- rrdhost_aclk_state_unlock(host);
-
- rrdhost_flag_set(host, RRDHOST_FLAG_METADATA_CLAIMID |RRDHOST_FLAG_METADATA_UPDATE);
-
- rrdpush_send_claimed_id(host);
-
- return PARSER_RC_OK;
-}
-
-// ----------------------------------------------------------------------------
-
-void pluginsd_process_thread_cleanup(void *ptr) {
- PARSER *parser = (PARSER *)ptr;
-
- pluginsd_cleanup_v2(parser);
- pluginsd_host_define_cleanup(parser);
-
- rrd_collector_finished();
-
-#ifdef NETDATA_LOG_STREAM_RECEIVE
- if(parser->user.stream_log_fp) {
- fclose(parser->user.stream_log_fp);
- parser->user.stream_log_fp = NULL;
- }
-#endif
-
- parser_destroy(parser);
-}
-
-bool parser_reconstruct_node(BUFFER *wb, void *ptr) {
- PARSER *parser = ptr;
- if(!parser || !parser->user.host)
- return false;
-
- buffer_strcat(wb, rrdhost_hostname(parser->user.host));
- return true;
-}
-
-bool parser_reconstruct_instance(BUFFER *wb, void *ptr) {
- PARSER *parser = ptr;
- if(!parser || !parser->user.st)
- return false;
-
- buffer_strcat(wb, rrdset_name(parser->user.st));
- return true;
-}
-
-bool parser_reconstruct_context(BUFFER *wb, void *ptr) {
- PARSER *parser = ptr;
- if(!parser || !parser->user.st)
- return false;
-
- buffer_strcat(wb, string2str(parser->user.st->context));
- return true;
-}
-
-inline size_t pluginsd_process(RRDHOST *host, struct plugind *cd, FILE *fp_plugin_input, FILE *fp_plugin_output, int trust_durations)
-{
- int enabled = cd->unsafe.enabled;
-
- if (!fp_plugin_input || !fp_plugin_output || !enabled) {
- cd->unsafe.enabled = 0;
- return 0;
- }
-
- if (unlikely(fileno(fp_plugin_input) == -1)) {
- netdata_log_error("input file descriptor given is not a valid stream");
- cd->serial_failures++;
- return 0;
- }
-
- if (unlikely(fileno(fp_plugin_output) == -1)) {
- netdata_log_error("output file descriptor given is not a valid stream");
- cd->serial_failures++;
- return 0;
- }
-
- clearerr(fp_plugin_input);
- clearerr(fp_plugin_output);
-
- PARSER *parser;
- {
- PARSER_USER_OBJECT user = {
- .enabled = cd->unsafe.enabled,
- .host = host,
- .cd = cd,
- .trust_durations = trust_durations
- };
-
- // fp_plugin_output = our input; fp_plugin_input = our output
- parser = parser_init(&user, fp_plugin_output, fp_plugin_input, -1, PARSER_INPUT_SPLIT, NULL);
- }
-
- pluginsd_keywords_init(parser, PARSER_INIT_PLUGINSD);
-
- rrd_collector_started();
-
- size_t count = 0;
-
- // this keeps the parser with its current value
- // so, parser needs to be allocated before pushing it
- netdata_thread_cleanup_push(pluginsd_process_thread_cleanup, parser);
-
- {
- ND_LOG_STACK lgs[] = {
- ND_LOG_FIELD_CB(NDF_REQUEST, line_splitter_reconstruct_line, &parser->line),
- ND_LOG_FIELD_CB(NDF_NIDL_NODE, parser_reconstruct_node, parser),
- ND_LOG_FIELD_CB(NDF_NIDL_INSTANCE, parser_reconstruct_instance, parser),
- ND_LOG_FIELD_CB(NDF_NIDL_CONTEXT, parser_reconstruct_context, parser),
- ND_LOG_FIELD_END(),
- };
- ND_LOG_STACK_PUSH(lgs);
-
- buffered_reader_init(&parser->reader);
- BUFFER *buffer = buffer_create(sizeof(parser->reader.read_buffer) + 2, NULL);
- while(likely(service_running(SERVICE_COLLECTORS))) {
-
- if(unlikely(!buffered_reader_next_line(&parser->reader, buffer))) {
- buffered_reader_ret_t ret = buffered_reader_read_timeout(
- &parser->reader,
- fileno((FILE *) parser->fp_input),
- 2 * 60 * MSEC_PER_SEC, true
- );
-
- if(unlikely(ret != BUFFERED_READER_READ_OK))
- break;
-
- continue;
- }
-
- if(unlikely(parser_action(parser, buffer->buffer)))
- break;
-
- buffer->len = 0;
- buffer->buffer[0] = '\0';
- }
- buffer_free(buffer);
-
- cd->unsafe.enabled = parser->user.enabled;
- count = parser->user.data_collections_count;
-
- if(likely(count)) {
- cd->successful_collections += count;
- cd->serial_failures = 0;
- }
- else
- cd->serial_failures++;
- }
-
- // free parser with the pop function
- netdata_thread_cleanup_pop(1);
-
- return count;
-}
-
-void pluginsd_keywords_init(PARSER *parser, PARSER_REPERTOIRE repertoire) {
- parser_init_repertoire(parser, repertoire);
-
- if (repertoire & (PARSER_INIT_PLUGINSD | PARSER_INIT_STREAMING))
- inflight_functions_init(parser);
-}
-
-PARSER *parser_init(struct parser_user_object *user, FILE *fp_input, FILE *fp_output, int fd,
- PARSER_INPUT_TYPE flags, void *ssl __maybe_unused) {
- PARSER *parser;
-
- parser = callocz(1, sizeof(*parser));
- if(user)
- parser->user = *user;
- parser->fd = fd;
- parser->fp_input = fp_input;
- parser->fp_output = fp_output;
-#ifdef ENABLE_HTTPS
- parser->ssl_output = ssl;
-#endif
- parser->flags = flags;
-
- spinlock_init(&parser->writer.spinlock);
- return parser;
-}
-
-PARSER_RC parser_execute(PARSER *parser, PARSER_KEYWORD *keyword, char **words, size_t num_words) {
- switch(keyword->id) {
- case 1:
- return pluginsd_set_v2(words, num_words, parser);
-
- case 2:
- return pluginsd_begin_v2(words, num_words, parser);
-
- case 3:
- return pluginsd_end_v2(words, num_words, parser);
-
- case 11:
- return pluginsd_set(words, num_words, parser);
-
- case 12:
- return pluginsd_begin(words, num_words, parser);
-
- case 13:
- return pluginsd_end(words, num_words, parser);
-
- case 21:
- return pluginsd_replay_set(words, num_words, parser);
-
- case 22:
- return pluginsd_replay_begin(words, num_words, parser);
-
- case 23:
- return pluginsd_replay_rrddim_collection_state(words, num_words, parser);
-
- case 24:
- return pluginsd_replay_rrdset_collection_state(words, num_words, parser);
-
- case 25:
- return pluginsd_replay_end(words, num_words, parser);
-
- case 31:
- return pluginsd_dimension(words, num_words, parser);
-
- case 32:
- return pluginsd_chart(words, num_words, parser);
-
- case 33:
- return pluginsd_chart_definition_end(words, num_words, parser);
-
- case 34:
- return pluginsd_clabel(words, num_words, parser);
-
- case 35:
- return pluginsd_clabel_commit(words, num_words, parser);
-
- case 41:
- return pluginsd_function(words, num_words, parser);
-
- case 42:
- return pluginsd_function_result_begin(words, num_words, parser);
-
- case 51:
- return pluginsd_label(words, num_words, parser);
-
- case 52:
- return pluginsd_overwrite(words, num_words, parser);
-
- case 53:
- return pluginsd_variable(words, num_words, parser);
-
- case 61:
- return streaming_claimed_id(words, num_words, parser);
-
- case 71:
- return pluginsd_host(words, num_words, parser);
-
- case 72:
- return pluginsd_host_define(words, num_words, parser);
-
- case 73:
- return pluginsd_host_define_end(words, num_words, parser);
-
- case 74:
- return pluginsd_host_labels(words, num_words, parser);
-
- case 97:
- return pluginsd_flush(words, num_words, parser);
-
- case 98:
- return pluginsd_disable(words, num_words, parser);
-
- case 99:
- return pluginsd_exit(words, num_words, parser);
-
- case 101:
- return pluginsd_register_plugin(words, num_words, parser);
-
- case 102:
- return pluginsd_register_module(words, num_words, parser);
-
- case 103:
- return pluginsd_register_job(words, num_words, parser);
-
- case 104:
- return pluginsd_dyncfg_reset(words, num_words, parser);
-
- case 110:
- return pluginsd_job_status(words, num_words, parser);
-
- case 111:
- return pluginsd_delete_job(words, num_words, parser);
-
- default:
- fatal("Unknown keyword '%s' with id %zu", keyword->keyword, keyword->id);
- }
-}
-
-#include "gperf-hashtable.h"
-
-void parser_init_repertoire(PARSER *parser, PARSER_REPERTOIRE repertoire) {
- parser->repertoire = repertoire;
-
- for(size_t i = GPERF_PARSER_MIN_HASH_VALUE ; i <= GPERF_PARSER_MAX_HASH_VALUE ;i++) {
- if(gperf_keywords[i].keyword && *gperf_keywords[i].keyword && (parser->repertoire & gperf_keywords[i].repertoire))
- worker_register_job_name(gperf_keywords[i].worker_job_id, gperf_keywords[i].keyword);
- }
-}
-
-static void parser_destroy_dyncfg(PARSER *parser) {
- if (parser->user.cd != NULL && parser->user.cd->configuration != NULL) {
- unregister_plugin(parser->user.host->configurable_plugins, parser->user.cd->cfg_dict_item);
- parser->user.cd->configuration = NULL;
- } else if (parser->user.host != NULL && SERVING_STREAMING(parser) && parser->user.host != localhost){
- dictionary_flush(parser->user.host->configurable_plugins);
- }
-}
-
-void parser_destroy(PARSER *parser) {
- if (unlikely(!parser))
- return;
-
- parser_destroy_dyncfg(parser);
-
- dictionary_destroy(parser->inflight.functions);
- freez(parser);
-}
-
-int pluginsd_parser_unittest(void) {
- PARSER *p = parser_init(NULL, NULL, NULL, -1, PARSER_INPUT_SPLIT, NULL);
- pluginsd_keywords_init(p, PARSER_INIT_PLUGINSD | PARSER_INIT_STREAMING);
-
- char *lines[] = {
- "BEGIN2 abcdefghijklmnopqr 123",
- "SET2 abcdefg 0x12345678 0 0",
- "SET2 hijklmnoqr 0x12345678 0 0",
- "SET2 stuvwxyz 0x12345678 0 0",
- "END2",
- NULL,
- };
-
- char *words[PLUGINSD_MAX_WORDS];
- size_t iterations = 1000000;
- size_t count = 0;
- char input[PLUGINSD_LINE_MAX + 1];
-
- usec_t started = now_realtime_usec();
- while(--iterations) {
- for(size_t line = 0; lines[line] ;line++) {
- strncpyz(input, lines[line], PLUGINSD_LINE_MAX);
- size_t num_words = quoted_strings_splitter_pluginsd(input, words, PLUGINSD_MAX_WORDS);
- const char *command = get_word(words, num_words, 0);
- PARSER_KEYWORD *keyword = parser_find_keyword(p, command);
- if(unlikely(!keyword))
- fatal("Cannot parse the line '%s'", lines[line]);
- count++;
- }
- }
- usec_t ended = now_realtime_usec();
-
- netdata_log_info("Parsed %zu lines in %0.2f secs, %0.2f klines/sec", count,
- (double)(ended - started) / (double)USEC_PER_SEC,
- (double)count / ((double)(ended - started) / (double)USEC_PER_SEC) / 1000.0);
-
- parser_destroy(p);
- return 0;
-}
diff --git a/collectors/proc.plugin/Makefile.am b/collectors/proc.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/proc.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/proc.plugin/README.md b/collectors/proc.plugin/README.md
deleted file mode 100644
index 62e46569f..000000000
--- a/collectors/proc.plugin/README.md
+++ /dev/null
@@ -1,639 +0,0 @@
-# OS provided metrics (proc.plugin)
-
-`proc.plugin` gathers metrics from the /proc and /sys folders in Linux systems, along with a few other endpoints, and is responsible for the bulk of the system metrics collected and visualized by Netdata.
-
-This plugin is not an external plugin, but one of Netdata's threads.
-
-In detail, it collects metrics from:
-
-- `/proc/net/dev` (all network interfaces for all their values)
-- `/proc/diskstats` (all disks for all their values)
-- `/proc/mdstat` (status of RAID arrays)
-- `/proc/net/snmp` (total IPv4, TCP and UDP usage)
-- `/proc/net/snmp6` (total IPv6 usage)
-- `/proc/net/netstat` (more IPv4 usage)
-- `/proc/net/wireless` (wireless extension)
-- `/proc/net/stat/nf_conntrack` (connection tracking performance)
-- `/proc/net/stat/synproxy` (synproxy performance)
-- `/proc/net/ip_vs/stats` (IPVS connection statistics)
-- `/proc/stat` (CPU utilization and attributes)
-- `/proc/meminfo` (memory information)
-- `/proc/vmstat` (system performance)
-- `/proc/net/rpc/nfsd` (NFS server statistics for both v3 and v4 NFS servers)
-- `/sys/fs/cgroup` (Control Groups - Linux Containers)
-- `/proc/self/mountinfo` (mount points)
-- `/proc/interrupts` (total and per core hardware interrupts)
-- `/proc/softirqs` (total and per core software interrupts)
-- `/proc/loadavg` (system load and total processes running)
-- `/proc/pressure/{cpu,memory,io}` (pressure stall information)
-- `/proc/sys/kernel/random/entropy_avail` (random numbers pool availability - used in cryptography)
-- `/proc/spl/kstat/zfs/arcstats` (status of ZFS adaptive replacement cache)
-- `/proc/spl/kstat/zfs/pool/state` (state of ZFS pools)
-- `/sys/class/power_supply` (power supply properties)
-- `/sys/class/infiniband` (infiniband interconnect)
-- `/sys/class/drm` (AMD GPUs)
-- `ipc` (IPC semaphores and message queues)
-- `ksm` Kernel Same-Page Merging performance (several files under `/sys/kernel/mm/ksm`).
-- `netdata` (internal Netdata resources utilization)
-
-- - -
-
-## Monitoring Disks
-
-> Live demo of disk monitoring at: **[http://london.netdata.rocks](https://registry.my-netdata.io/#menu_disk)**
-
-Performance monitoring for Linux disks is quite complicated. The main reason is the plethora of disk technologies available. There are many different hardware disk technologies, but there are even more **virtual disk** technologies that can provide additional storage features.
-
-Hopefully, the Linux kernel provides many metrics that can provide deep insights of what our disks our doing. The kernel measures all these metrics on all layers of storage: **virtual disks**, **physical disks** and **partitions of disks**.
-
-### Monitored disk metrics
-
-- **I/O bandwidth/s (kb/s)**
- The amount of data transferred from and to the disk.
-- **Amount of discarded data (kb/s)**
-- **I/O operations/s**
- The number of I/O operations completed.
-- **Extended I/O operations/s**
- The number of extended I/O operations completed.
-- **Queued I/O operations**
- The number of currently queued I/O operations. For traditional disks that execute commands one after another, one of them is being run by the disk and the rest are just waiting in a queue.
-- **Backlog size (time in ms)**
- The expected duration of the currently queued I/O operations.
-- **Utilization (time percentage)**
- The percentage of time the disk was busy with something. This is a very interesting metric, since for most disks, that execute commands sequentially, **this is the key indication of congestion**. A sequential disk that is 100% of the available time busy, has no time to do anything more, so even if the bandwidth or the number of operations executed by the disk is low, its capacity has been reached.
- Of course, for newer disk technologies (like fusion cards) that are capable to execute multiple commands in parallel, this metric is just meaningless.
-- **Average I/O operation time (ms)**
- The average time for I/O requests issued to the device to be served. This includes the time spent by the requests in queue and the time spent servicing them.
-- **Average I/O operation time for extended operations (ms)**
- The average time for extended I/O requests issued to the device to be served. This includes the time spent by the requests in queue and the time spent servicing them.
-- **Average I/O operation size (kb)**
- The average amount of data of the completed I/O operations.
-- **Average amount of discarded data (kb)**
- The average amount of data of the completed discard operations.
-- **Average Service Time (ms)**
- The average service time for completed I/O operations. This metric is calculated using the total busy time of the disk and the number of completed operations. If the disk is able to execute multiple parallel operations the reporting average service time will be misleading.
-- **Average Service Time for extended I/O operations (ms)**
- The average service time for completed extended I/O operations.
-- **Merged I/O operations/s**
- The Linux kernel is capable of merging I/O operations. So, if two requests to read data from the disk are adjacent, the Linux kernel may merge them to one before giving them to disk. This metric measures the number of operations that have been merged by the Linux kernel.
-- **Merged discard operations/s**
-- **Total I/O time**
- The sum of the duration of all completed I/O operations. This number can exceed the interval if the disk is able to execute multiple I/O operations in parallel.
-- **Space usage**
- For mounted disks, Netdata will provide a chart for their space, with 3 dimensions:
- 1. free
- 2. used
- 3. reserved for root
-- **inode usage**
- For mounted disks, Netdata will provide a chart for their inodes (number of file and directories), with 3 dimensions:
- 1. free
- 2. used
- 3. reserved for root
-
-### disk names
-
-Netdata will automatically set the name of disks on the dashboard, from the mount point they are mounted, of course only when they are mounted. Changes in mount points are not currently detected (you will have to restart Netdata to change the name of the disk). To use disk IDs provided by `/dev/disk/by-id`, the `name disks by id` option should be enabled. The `preferred disk ids` simple pattern allows choosing disk IDs to be used in the first place.
-
-### performance metrics
-
-By default, Netdata will enable monitoring metrics only when they are not zero. If they are constantly zero they are ignored. Metrics that will start having values, after Netdata is started, will be detected and charts will be automatically added to the dashboard (a refresh of the dashboard is needed for them to appear though). Set `yes` for a chart instead of `auto` to enable it permanently. You can also set the `enable zero metrics` option to `yes` in the `[global]` section which enables charts with zero metrics for all internal Netdata plugins.
-
-Netdata categorizes all block devices in 3 categories:
-
-1. physical disks (i.e. block devices that do not have child devices and are not partitions)
-2. virtual disks (i.e. block devices that have child devices - like RAID devices)
-3. disk partitions (i.e. block devices that are part of a physical disk)
-
-Performance metrics are enabled by default for all disk devices, except partitions and not-mounted virtual disks. Of course, you can enable/disable monitoring any block device by editing the Netdata configuration file.
-
-### Netdata configuration
-
-You can get the running Netdata configuration using this:
-
-```sh
-cd /etc/netdata
-curl "http://localhost:19999/netdata.conf" >netdata.conf.new
-mv netdata.conf.new netdata.conf
-```
-
-Then edit `netdata.conf` and find the following section. This is the basic plugin configuration.
-
-```
-[plugin:proc:/proc/diskstats]
- # enable new disks detected at runtime = yes
- # performance metrics for physical disks = auto
- # performance metrics for virtual disks = auto
- # performance metrics for partitions = no
- # bandwidth for all disks = auto
- # operations for all disks = auto
- # merged operations for all disks = auto
- # i/o time for all disks = auto
- # queued operations for all disks = auto
- # utilization percentage for all disks = auto
- # extended operations for all disks = auto
- # backlog for all disks = auto
- # bcache for all disks = auto
- # bcache priority stats update every = 0
- # remove charts of removed disks = yes
- # path to get block device = /sys/block/%s
- # path to get block device bcache = /sys/block/%s/bcache
- # path to get virtual block device = /sys/devices/virtual/block/%s
- # path to get block device infos = /sys/dev/block/%lu:%lu/%s
- # path to device mapper = /dev/mapper
- # path to /dev/disk/by-label = /dev/disk/by-label
- # path to /dev/disk/by-id = /dev/disk/by-id
- # path to /dev/vx/dsk = /dev/vx/dsk
- # name disks by id = no
- # preferred disk ids = *
- # exclude disks = loop* ram*
- # filename to monitor = /proc/diskstats
- # performance metrics for disks with major 8 = yes
-```
-
-For each virtual disk, physical disk and partition you will have a section like this:
-
-```
-[plugin:proc:/proc/diskstats:sda]
- # enable = yes
- # enable performance metrics = auto
- # bandwidth = auto
- # operations = auto
- # merged operations = auto
- # i/o time = auto
- # queued operations = auto
- # utilization percentage = auto
- # extended operations = auto
- # backlog = auto
-```
-
-For all configuration options:
-
-- `auto` = enable monitoring if the collected values are not zero
-- `yes` = enable monitoring
-- `no` = disable monitoring
-
-Of course, to set options, you will have to uncomment them. The comments show the internal defaults.
-
-After saving `/etc/netdata/netdata.conf`, restart your Netdata to apply them.
-
-#### Disabling performance metrics for individual device and to multiple devices by device type
-
-You can pretty easy disable performance metrics for individual device, for ex.:
-
-```
-[plugin:proc:/proc/diskstats:sda]
- enable performance metrics = no
-```
-
-But sometimes you need disable performance metrics for all devices with the same type, to do it you need to figure out device type from `/proc/diskstats` for ex.:
-
-```
- 7 0 loop0 1651 0 3452 168 0 0 0 0 0 8 168
- 7 1 loop1 4955 0 11924 880 0 0 0 0 0 64 880
- 7 2 loop2 36 0 216 4 0 0 0 0 0 4 4
- 7 6 loop6 0 0 0 0 0 0 0 0 0 0 0
- 7 7 loop7 0 0 0 0 0 0 0 0 0 0 0
- 251 2 zram2 27487 0 219896 188 79953 0 639624 1640 0 1828 1828
- 251 3 zram3 27348 0 218784 152 79952 0 639616 1960 0 2060 2104
-```
-
-All zram devices starts with `251` number and all loop devices starts with `7`.
-So, to disable performance metrics for all loop devices you could add `performance metrics for disks with major 7 = no` to `[plugin:proc:/proc/diskstats]` section.
-
-```
-[plugin:proc:/proc/diskstats]
- performance metrics for disks with major 7 = no
-```
-
-## Monitoring RAID arrays
-
-### Monitored RAID array metrics
-
-1. **Health** Number of failed disks in every array (aggregate chart).
-
-2. **Disks stats**
-
-- total (number of devices array ideally would have)
-- inuse (number of devices currently are in use)
-
-3. **Mismatch count**
-
-- unsynchronized blocks
-
-4. **Current status**
-
-- resync in percent
-- recovery in percent
-- reshape in percent
-- check in percent
-
-5. **Operation status** (if resync/recovery/reshape/check is active)
-
-- finish in minutes
-- speed in megabytes/s
-
-6. **Nonredundant array availability**
-
-#### configuration
-
-```
-[plugin:proc:/proc/mdstat]
- # faulty devices = yes
- # nonredundant arrays availability = yes
- # mismatch count = auto
- # disk stats = yes
- # operation status = yes
- # make charts obsolete = yes
- # filename to monitor = /proc/mdstat
- # mismatch_cnt filename to monitor = /sys/block/%s/md/mismatch_cnt
-```
-
-## Monitoring CPUs
-
-The `/proc/stat` module monitors CPU utilization, interrupts, context switches, processes started/running, thermal
-throttling, frequency, and idle states. It gathers this information from multiple files.
-
-If your system has more than 50 processors (`physical processors * cores per processor * threads per core`), the Agent
-automatically disables CPU thermal throttling, frequency, and idle state charts. To override this default, see the next
-section on configuration.
-
-### Configuration
-
-The settings for monitoring CPUs is in the `[plugin:proc:/proc/stat]` of your `netdata.conf` file.
-
-The `keep per core files open` option lets you reduce the number of file operations on multiple files.
-
-If your system has more than 50 processors and you would like to see the CPU thermal throttling, frequency, and idle
-state charts that are automatically disabled, you can set the following boolean options in the
-`[plugin:proc:/proc/stat]` section.
-
-```conf
- keep per core files open = yes
- keep cpuidle files open = yes
- core_throttle_count = yes
- package_throttle_count = yes
- cpu frequency = yes
- cpu idle states = yes
-```
-
-### CPU frequency
-
-The module shows the current CPU frequency as set by the `cpufreq` kernel
-module.
-
-**Requirement:**
-You need to have `CONFIG_CPU_FREQ` and (optionally) `CONFIG_CPU_FREQ_STAT`
-enabled in your kernel.
-
-`cpufreq` interface provides two different ways of getting the information through `/sys/devices/system/cpu/cpu*/cpufreq/scaling_cur_freq` and `/sys/devices/system/cpu/cpu*/cpufreq/stats/time_in_state` files. The latter is more accurate so it is preferred in the module. `scaling_cur_freq` represents only the current CPU frequency, and doesn't account for any state changes which happen between updates. The module switches back and forth between these two methods if governor is changed.
-
-It produces one chart with multiple lines (one line per core).
-
-#### configuration
-
-`scaling_cur_freq filename to monitor` and `time_in_state filename to monitor` in the `[plugin:proc:/proc/stat]` configuration section
-
-### CPU idle states
-
-The module monitors the usage of CPU idle states.
-
-**Requirement:**
-Your kernel needs to have `CONFIG_CPU_IDLE` enabled.
-
-It produces one stacked chart per CPU, showing the percentage of time spent in
-each state.
-
-#### configuration
-
-`schedstat filename to monitor`, `cpuidle name filename to monitor`, and `cpuidle time filename to monitor` in the `[plugin:proc:/proc/stat]` configuration section
-
-## Monitoring memory
-
-### Monitored memory metrics
-
-- Amount of memory swapped in/out
-- Amount of memory paged from/to disk
-- Number of memory page faults
-- Number of out of memory kills
-- Number of NUMA events
-
-### Configuration
-
-```conf
-[plugin:proc:/proc/vmstat]
- filename to monitor = /proc/vmstat
- swap i/o = auto
- disk i/o = yes
- memory page faults = yes
- out of memory kills = yes
- system-wide numa metric summary = auto
-```
-
-## Monitoring Network Interfaces
-
-### Monitored network interface metrics
-
-- **Physical Network Interfaces Aggregated Bandwidth (kilobits/s)**
- The amount of data received and sent through all physical interfaces in the system. This is the source of data for the Net Inbound and Net Outbound dials in the System Overview section.
-
-- **Bandwidth (kilobits/s)**
- The amount of data received and sent through the interface.
-
-- **Packets (packets/s)**
- The number of packets received, packets sent, and multicast packets transmitted through the interface.
-
-- **Interface Errors (errors/s)**
- The number of errors for the inbound and outbound traffic on the interface.
-
-- **Interface Drops (drops/s)**
- The number of packets dropped for the inbound and outbound traffic on the interface.
-
-- **Interface FIFO Buffer Errors (errors/s)**
- The number of FIFO buffer errors encountered while receiving and transmitting data through the interface.
-
-- **Compressed Packets (packets/s)**
- The number of compressed packets transmitted or received by the device driver.
-
-- **Network Interface Events (events/s)**
- The number of packet framing errors, collisions detected on the interface, and carrier losses detected by the device driver.
-
-By default Netdata will enable monitoring metrics only when they are not zero. If they are constantly zero they are ignored. Metrics that will start having values, after Netdata is started, will be detected and charts will be automatically added to the dashboard (a refresh of the dashboard is needed for them to appear though).
-
-### Monitoring wireless network interfaces
-
-The settings for monitoring wireless is in the `[plugin:proc:/proc/net/wireless]` section of your `netdata.conf` file.
-
-```conf
- status for all interfaces = yes
- quality for all interfaces = yes
- discarded packets for all interfaces = yes
- missed beacon for all interface = yes
-```
-
-You can set the following values for each configuration option:
-
-- `auto` = enable monitoring if the collected values are not zero
-- `yes` = enable monitoring
-- `no` = disable monitoring
-
-#### Monitored wireless interface metrics
-
-- **Status**
- The current state of the interface. This is a device-dependent option.
-
-- **Link**
- Overall quality of the link.
-
-- **Level**
- Received signal strength (RSSI), which indicates how strong the received signal is.
-
-- **Noise**
- Background noise level.
-
-- **Discarded packets**
- Discarded packets for: Number of packets received with a different NWID or ESSID (`nwid`), unable to decrypt (`crypt`), hardware was not able to properly re-assemble the link layer fragments (`frag`), packets failed to deliver (`retry`), and packets lost in relation with specific wireless operations (`misc`).
-
-- **Missed beacon**
- Number of periodic beacons from the cell or the access point the interface has missed.
-
-#### Wireless configuration
-
-#### alerts
-
-There are several alerts defined in `health.d/net.conf`.
-
-The tricky ones are `inbound packets dropped` and `inbound packets dropped ratio`. They have quite a strict policy so that they warn users about possible issues. These alerts can be annoying for some network configurations. It is especially true for some bonding configurations if an interface is a child or a bonding interface itself. If it is expected to have a certain number of drops on an interface for a certain network configuration, a separate alert with different triggering thresholds can be created or the existing one can be disabled for this specific interface. It can be done with the help of the [families](https://github.com/netdata/netdata/blob/master/health/REFERENCE.md#alert-line-families) line in the alert configuration. For example, if you want to disable the `inbound packets dropped` alert for `eth0`, set `families: !eth0 *` in the alert definition for `template: inbound_packets_dropped`.
-
-#### configuration
-
-Module configuration:
-
-```
-[plugin:proc:/proc/net/dev]
- # filename to monitor = /proc/net/dev
- # path to get virtual interfaces = /sys/devices/virtual/net/%s
- # path to get net device speed = /sys/class/net/%s/speed
- # enable new interfaces detected at runtime = auto
- # bandwidth for all interfaces = auto
- # packets for all interfaces = auto
- # errors for all interfaces = auto
- # drops for all interfaces = auto
- # fifo for all interfaces = auto
- # compressed packets for all interfaces = auto
- # frames, collisions, carrier counters for all interfaces = auto
- # disable by default interfaces matching = lo fireqos* *-ifb
- # refresh interface speed every seconds = 10
-```
-
-Per interface configuration:
-
-```
-[plugin:proc:/proc/net/dev:enp0s3]
- # enabled = yes
- # virtual = no
- # bandwidth = auto
- # packets = auto
- # errors = auto
- # drops = auto
- # fifo = auto
- # compressed = auto
- # events = auto
-```
-
-## Linux Anti-DDoS
-
-![image6](https://cloud.githubusercontent.com/assets/2662304/14253733/53550b16-fa95-11e5-8d9d-4ed171df4735.gif)
-
----
-
-SYNPROXY is a TCP SYN packets proxy. It can be used to protect any TCP server (like a web server) from SYN floods and similar DDos attacks.
-
-SYNPROXY is a netfilter module, in the Linux kernel (since version 3.12). It is optimized to handle millions of packets per second utilizing all CPUs available without any concurrency locking between the connections.
-
-The net effect of this, is that the real servers will not notice any change during the attack. The valid TCP connections will pass through and served, while the attack will be stopped at the firewall.
-
-Netdata does not enable SYNPROXY. It just uses the SYNPROXY metrics exposed by your kernel, so you will first need to configure it. The hard way is to run iptables SYNPROXY commands directly on the console. An easier way is to use [FireHOL](https://firehol.org/), which, is a firewall manager for iptables. FireHOL can configure SYNPROXY using the following setup guides:
-
-- **[Working with SYNPROXY](https://github.com/firehol/firehol/wiki/Working-with-SYNPROXY)**
-- **[Working with SYNPROXY and traps](https://github.com/firehol/firehol/wiki/Working-with-SYNPROXY-and-traps)**
-
-### Real-time monitoring of Linux Anti-DDoS
-
-Netdata is able to monitor in real-time (per second updates) the operation of the Linux Anti-DDoS protection.
-
-It visualizes 4 charts:
-
-1. TCP SYN Packets received on ports operated by SYNPROXY
-2. TCP Cookies (valid, invalid, retransmits)
-3. Connections Reopened
-4. Entries used
-
-Example image:
-
-![ddos](https://cloud.githubusercontent.com/assets/2662304/14398891/6016e3fc-fdf0-11e5-942b-55de6a52cb66.gif)
-
-See Linux Anti-DDoS in action at: **[Netdata demo site (with SYNPROXY enabled)](https://registry.my-netdata.io/#menu_netfilter_submenu_synproxy)**
-
-## Linux power supply
-
-This module monitors various metrics reported by power supply drivers
-on Linux. This allows tracking and alerting on things like remaining
-battery capacity.
-
-Depending on the underlying driver, it may provide the following charts
-and metrics:
-
-1. Capacity: The power supply capacity expressed as a percentage.
-
- - capacity_now
-
-2. Charge: The charge for the power supply, expressed as amphours.
-
- - charge_full_design
- - charge_full
- - charge_now
- - charge_empty
- - charge_empty_design
-
-3. Energy: The energy for the power supply, expressed as watthours.
-
- - energy_full_design
- - energy_full
- - energy_now
- - energy_empty
- - energy_empty_design
-
-4. Voltage: The voltage for the power supply, expressed as volts.
-
- - voltage_max_design
- - voltage_max
- - voltage_now
- - voltage_min
- - voltage_min_design
-
-#### configuration
-
-```
-[plugin:proc:/sys/class/power_supply]
- # battery capacity = yes
- # battery charge = no
- # battery energy = no
- # power supply voltage = no
- # keep files open = auto
- # directory to monitor = /sys/class/power_supply
-```
-
-#### notes
-
-- Most drivers provide at least the first chart. Battery powered ACPI
- compliant systems (like most laptops) provide all but the third, but do
- not provide all of the metrics for each chart.
-
-- Current, energy, and voltages are reported with a *very* high precision
- by the power_supply framework. Usually, this is far higher than the
- actual hardware supports reporting, so expect to see changes in these
- charts jump instead of scaling smoothly.
-
-- If `max` or `full` attribute is defined by the driver, but not a
- corresponding `min` or `empty` attribute, then Netdata will still provide
- the corresponding `min` or `empty`, which will then always read as zero.
- This way, alerts which match on these will still work.
-
-## Infiniband interconnect
-
-This module monitors every active Infiniband port. It provides generic counters statistics, and per-vendor hw-counters (if vendor is supported).
-
-### Monitored interface metrics
-
-Each port will have its counters metrics monitored, grouped in the following charts:
-
-- **Bandwidth usage**
- Sent/Received data, in KB/s
-
-- **Packets Statistics**
- Sent/Received packets, in 3 categories: total, unicast and multicast.
-
-- **Errors Statistics**
- Many errors counters are provided, presenting statistics for:
- - Packets: malformed, sent/received discarded by card/switch, missing resource
- - Link: downed, recovered, integrity error, minor error
- - Other events: Tick Wait to send, buffer overrun
-
-If your vendor is supported, you'll also get HW-Counters statistics. These being vendor specific, please refer to their documentation.
-
-- Mellanox: [see statistics documentation](https://community.mellanox.com/s/article/understanding-mlx5-linux-counters-and-status-parameters)
-
-### configuration
-
-Default configuration will monitor only enabled infiniband ports, and refresh newly activated or created ports every 30 seconds
-
-```
-[plugin:proc:/sys/class/infiniband]
- # dirname to monitor = /sys/class/infiniband
- # bandwidth counters = yes
- # packets counters = yes
- # errors counters = yes
- # hardware packets counters = auto
- # hardware errors counters = auto
- # monitor only ports being active = auto
- # disable by default interfaces matching =
- # refresh ports state every seconds = 30
-```
-
-## AMD GPUs
-
-This module monitors every AMD GPU card discovered at agent startup.
-
-### Monitored GPU metrics
-
-The following charts will be provided:
-
-- **GPU utilization**
-- **GPU memory utilization**
-- **GPU clock frequency**
-- **GPU memory clock frequency**
-- **VRAM memory usage percentage**
-- **VRAM memory usage**
-- **visible VRAM memory usage percentage**
-- **visible VRAM memory usage**
-- **GTT memory usage percentage**
-- **GTT memory usage**
-
-### configuration
-
-The `drm` path can be configured if it differs from the default:
-
-```
-[plugin:proc:/sys/class/drm]
- # directory to monitor = /sys/class/drm
-```
-
-> [!NOTE]
-> Temperature, fan speed, voltage and power metrics for AMD GPUs can be monitored using the [Sensors](https://github.com/netdata/netdata/blob/master/collectors/charts.d.plugin/sensors/README.md) plugin.
-
-## IPC
-
-### Monitored IPC metrics
-
-- **number of messages in message queues**
-- **amount of memory used by message queues**
-- **number of semaphores**
-- **number of semaphore arrays**
-- **number of shared memory segments**
-- **amount of memory used by shared memory segments**
-
-As far as the message queue charts are dynamic, sane limits are applied for the number of dimensions per chart (the limit is configurable).
-
-### configuration
-
-```
-[plugin:proc:ipc]
- # message queues = yes
- # semaphore totals = yes
- # shared memory totals = yes
- # msg filename to monitor = /proc/sysvipc/msg
- # shm filename to monitor = /proc/sysvipc/shm
- # max dimensions in memory allowed = 50
-```
-
-
diff --git a/collectors/proc.plugin/integrations/btrfs.md b/collectors/proc.plugin/integrations/btrfs.md
deleted file mode 100644
index 5f994c841..000000000
--- a/collectors/proc.plugin/integrations/btrfs.md
+++ /dev/null
@@ -1,137 +0,0 @@
-<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/btrfs.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
-sidebar_label: "BTRFS"
-learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Filesystem/BTRFS"
-most_popular: False
-message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
-endmeta-->
-
-# BTRFS
-
-
-<img src="https://netdata.cloud/img/filesystem.svg" width="150"/>
-
-
-Plugin: proc.plugin
-Module: /sys/fs/btrfs
-
-<img src="https://img.shields.io/badge/maintained%20by-Netdata-%2300ab44" />
-
-## Overview
-
-This integration provides usage and error statistics from the BTRFS filesystem.
-
-
-
-This collector is supported on all platforms.
-
-This collector supports collecting metrics from multiple instances of this integration, including remote instances.
-
-
-### Default Behavior
-
-#### Auto-Detection
-
-This integration doesn't support auto-detection.
-
-#### Limits
-
-The default configuration for this integration does not impose any limits on data collection.
-
-#### Performance Impact
-
-The default configuration for this integration is not expected to impose a significant performance impact on the system.
-
-
-## Metrics
-
-Metrics grouped by *scope*.
-
-The scope defines the instance that the metric belongs to. An instance is uniquely identified by a set of labels.
-
-
-
-### Per btrfs filesystem
-
-
-
-Labels:
-
-| Label | Description |
-|:-----------|:----------------|
-| filesystem_uuid | TBD |
-| filesystem_label | TBD |
-
-Metrics:
-
-| Metric | Dimensions | Unit |
-|:------|:----------|:----|
-| btrfs.disk | unallocated, data_free, data_used, meta_free, meta_used, sys_free, sys_used | MiB |
-| btrfs.data | free, used | MiB |
-| btrfs.metadata | free, used, reserved | MiB |
-| btrfs.system | free, used | MiB |
-| btrfs.commits | commits | commits |
-| btrfs.commits_perc_time | commits | percentage |
-| btrfs.commit_timings | last, max | ms |
-
-### Per btrfs device
-
-
-
-Labels:
-
-| Label | Description |
-|:-----------|:----------------|
-| device_id | TBD |
-| filesystem_uuid | TBD |
-| filesystem_label | TBD |
-
-Metrics:
-
-| Metric | Dimensions | Unit |
-|:------|:----------|:----|
-| btrfs.device_errors | write_errs, read_errs, flush_errs, corruption_errs, generation_errs | errors |
-
-
-
-## Alerts
-
-
-The following alerts are available:
-
-| Alert name | On metric | Description |
-|:------------|:----------|:------------|
-| [ btrfs_allocated ](https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf) | btrfs.disk | percentage of allocated BTRFS physical disk space |
-| [ btrfs_data ](https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf) | btrfs.data | utilization of BTRFS data space |
-| [ btrfs_metadata ](https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf) | btrfs.metadata | utilization of BTRFS metadata space |
-| [ btrfs_system ](https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf) | btrfs.system | utilization of BTRFS system space |
-| [ btrfs_device_read_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf) | btrfs.device_errors | number of encountered BTRFS read errors |
-| [ btrfs_device_write_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf) | btrfs.device_errors | number of encountered BTRFS write errors |
-| [ btrfs_device_flush_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf) | btrfs.device_errors | number of encountered BTRFS flush errors |
-| [ btrfs_device_corruption_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf) | btrfs.device_errors | number of encountered BTRFS corruption errors |
-| [ btrfs_device_generation_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf) | btrfs.device_errors | number of encountered BTRFS generation errors |
-
-
-## Setup
-
-### Prerequisites
-
-No action required.
-
-### Configuration
-
-#### File
-
-There is no configuration file.
-#### Options
-
-
-
-There are no configuration options.
-
-#### Examples
-There are no configuration examples.
-
-
diff --git a/collectors/proc.plugin/integrations/network_interfaces.md b/collectors/proc.plugin/integrations/network_interfaces.md
deleted file mode 100644
index 0cfd56fae..000000000
--- a/collectors/proc.plugin/integrations/network_interfaces.md
+++ /dev/null
@@ -1,137 +0,0 @@
-<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/network_interfaces.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
-sidebar_label: "Network interfaces"
-learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
-most_popular: False
-message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
-endmeta-->
-
-# Network interfaces
-
-
-<img src="https://netdata.cloud/img/network-wired.svg" width="150"/>
-
-
-Plugin: proc.plugin
-Module: /proc/net/dev
-
-<img src="https://img.shields.io/badge/maintained%20by-Netdata-%2300ab44" />
-
-## Overview
-
-Monitor network interface metrics about bandwidth, state, errors and more.
-
-
-
-This collector is supported on all platforms.
-
-This collector supports collecting metrics from multiple instances of this integration, including remote instances.
-
-
-### Default Behavior
-
-#### Auto-Detection
-
-This integration doesn't support auto-detection.
-
-#### Limits
-
-The default configuration for this integration does not impose any limits on data collection.
-
-#### Performance Impact
-
-The default configuration for this integration is not expected to impose a significant performance impact on the system.
-
-
-## Metrics
-
-Metrics grouped by *scope*.
-
-The scope defines the instance that the metric belongs to. An instance is uniquely identified by a set of labels.
-
-
-
-### Per Network interfaces instance
-
-
-
-This scope has no labels.
-
-Metrics:
-
-| Metric | Dimensions | Unit |
-|:------|:----------|:----|
-| system.net | received, sent | kilobits/s |
-
-### Per network device
-
-
-
-Labels:
-
-| Label | Description |
-|:-----------|:----------------|
-| interface_type | TBD |
-| device | TBD |
-
-Metrics:
-
-| Metric | Dimensions | Unit |
-|:------|:----------|:----|
-| net.net | received, sent | kilobits/s |
-| net.speed | speed | kilobits/s |
-| net.duplex | full, half, unknown | state |
-| net.operstate | up, down, notpresent, lowerlayerdown, testing, dormant, unknown | state |
-| net.carrier | up, down | state |
-| net.mtu | mtu | octets |
-| net.packets | received, sent, multicast | packets/s |
-| net.errors | inbound, outbound | errors/s |
-| net.drops | inbound, outbound | drops/s |
-| net.fifo | receive, transmit | errors |
-| net.compressed | received, sent | packets/s |
-| net.events | frames, collisions, carrier | events/s |
-
-
-
-## Alerts
-
-
-The following alerts are available:
-
-| Alert name | On metric | Description |
-|:------------|:----------|:------------|
-| [ interface_speed ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.net | network interface ${label:device} current speed |
-| [ 1m_received_traffic_overflow ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.net | average inbound utilization for the network interface ${label:device} over the last minute |
-| [ 1m_sent_traffic_overflow ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.net | average outbound utilization for the network interface ${label:device} over the last minute |
-| [ inbound_packets_dropped_ratio ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.drops | ratio of inbound dropped packets for the network interface ${label:device} over the last 10 minutes |
-| [ outbound_packets_dropped_ratio ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.drops | ratio of outbound dropped packets for the network interface ${label:device} over the last 10 minutes |
-| [ wifi_inbound_packets_dropped_ratio ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.drops | ratio of inbound dropped packets for the network interface ${label:device} over the last 10 minutes |
-| [ wifi_outbound_packets_dropped_ratio ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.drops | ratio of outbound dropped packets for the network interface ${label:device} over the last 10 minutes |
-| [ 1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.packets | average number of packets received by the network interface ${label:device} over the last minute |
-| [ 10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
-| [ 10min_fifo_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.fifo | number of FIFO errors for the network interface ${label:device} in the last 10 minutes |
-
-
-## Setup
-
-### Prerequisites
-
-No action required.
-
-### Configuration
-
-#### File
-
-There is no configuration file.
-#### Options
-
-
-
-There are no configuration options.
-
-#### Examples
-There are no configuration examples.
-
-
diff --git a/collectors/proc.plugin/metadata.yaml b/collectors/proc.plugin/metadata.yaml
deleted file mode 100644
index 45351b36f..000000000
--- a/collectors/proc.plugin/metadata.yaml
+++ /dev/null
@@ -1,5299 +0,0 @@
-plugin_name: proc.plugin
-modules:
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/stat
- monitored_instance:
- name: System statistics
- link: ""
- categories:
- - data-collection.linux-systems.system-metrics
- icon_filename: "linuxserver.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - cpu utilization
- - process counts
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- CPU utilization, states and frequencies and key Linux system performance metrics.
-
- The `/proc/stat` file provides various types of system statistics:
-
- - The overall system CPU usage statistics
- - Per CPU core statistics
- - The total context switching of the system
- - The total number of processes running
- - The total CPU interrupts
- - The total CPU softirqs
-
- The collector also reads:
-
- - `/proc/schedstat` for statistics about the process scheduler in the Linux kernel.
- - `/sys/devices/system/cpu/[X]/thermal_throttle/core_throttle_count` to get the count of thermal throttling events for a specific CPU core on Linux systems.
- - `/sys/devices/system/cpu/[X]/thermal_throttle/package_throttle_count` to get the count of thermal throttling events for a specific CPU package on a Linux system.
- - `/sys/devices/system/cpu/[X]/cpufreq/scaling_cur_freq` to get the current operating frequency of a specific CPU core.
- - `/sys/devices/system/cpu/[X]/cpufreq/stats/time_in_state` to get the amount of time the CPU has spent in each of its available frequency states.
- - `/sys/devices/system/cpu/[X]/cpuidle/state[X]/name` to get the names of the idle states for each CPU core in a Linux system.
- - `/sys/devices/system/cpu/[X]/cpuidle/state[X]/time` to get the total time each specific CPU core has spent in each idle state since the system was started.
- method_description: ""
- supported_platforms:
- include: ["linux"]
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: |
- The collector auto-detects all metrics. No configuration is needed.
- limits:
- description: ""
- performance_impact:
- description: |
- The collector disables cpu frequency and idle state monitoring when there are more than 128 CPU cores available.
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- section_name: "plugin:proc:/proc/stat"
- name: "netdata.conf"
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 10min_cpu_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf
- metric: system.cpu
- info: average CPU utilization over the last 10 minutes (excluding iowait, nice and steal)
- os: "linux"
- - name: 10min_cpu_iowait
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf
- metric: system.cpu
- info: average CPU iowait time over the last 10 minutes
- os: "linux"
- - name: 20min_steal_cpu
- link: https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf
- metric: system.cpu
- info: average CPU steal time over the last 20 minutes
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.cpu
- description: Total CPU utilization
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: guest_nice
- - name: guest
- - name: steal
- - name: softirq
- - name: irq
- - name: user
- - name: system
- - name: nice
- - name: iowait
- - name: idle
- - name: system.intr
- description: CPU Interrupts
- unit: "interrupts/s"
- chart_type: line
- dimensions:
- - name: interrupts
- - name: system.ctxt
- description: CPU Context Switches
- unit: "context switches/s"
- chart_type: line
- dimensions:
- - name: switches
- - name: system.forks
- description: Started Processes
- unit: "processes/s"
- chart_type: line
- dimensions:
- - name: started
- - name: system.processes
- description: System Processes
- unit: "processes"
- chart_type: line
- dimensions:
- - name: running
- - name: blocked
- - name: cpu.core_throttling
- description: Core Thermal Throttling Events
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: a dimension per cpu core
- - name: cpu.package_throttling
- description: Package Thermal Throttling Events
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: a dimension per package
- - name: cpu.cpufreq
- description: Current CPU Frequency
- unit: "MHz"
- chart_type: line
- dimensions:
- - name: a dimension per cpu core
- - name: cpu core
- description: ""
- labels:
- - name: cpu
- description: TBD
- metrics:
- - name: cpu.cpu
- description: Core utilization
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: guest_nice
- - name: guest
- - name: steal
- - name: softirq
- - name: irq
- - name: user
- - name: system
- - name: nice
- - name: iowait
- - name: idle
- - name: cpuidle.cpu_cstate_residency_time
- description: C-state residency time
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: a dimension per c-state
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/sys/kernel/random/entropy_avail
- monitored_instance:
- name: Entropy
- link: ""
- categories:
- - data-collection.linux-systems.system-metrics
- icon_filename: "syslog.png"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - entropy
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- Entropy, a measure of the randomness or unpredictability of data.
-
- In the context of cryptography, entropy is used to generate random numbers or keys that are essential for
- secure communication and encryption. Without a good source of entropy, cryptographic protocols can become
- vulnerable to attacks that exploit the predictability of the generated keys.
-
- In most operating systems, entropy is generated by collecting random events from various sources, such as
- hardware interrupts, mouse movements, keyboard presses, and disk activity. These events are fed into a pool
- of entropy, which is then used to generate random numbers when needed.
-
- The `/dev/random` device in Linux is one such source of entropy, and it provides an interface for programs
- to access the pool of entropy. When a program requests random numbers, it reads from the `/dev/random` device,
- which blocks until enough entropy is available to generate the requested numbers. This ensures that the
- generated numbers are truly random and not predictable.
-
- However, if the pool of entropy gets depleted, the `/dev/random` device may block indefinitely, causing
- programs that rely on random numbers to slow down or even freeze. This is especially problematic for
- cryptographic protocols that require a continuous stream of random numbers, such as SSL/TLS and SSH.
-
- To avoid this issue, some systems use a hardware random number generator (RNG) to generate high-quality
- entropy. A hardware RNG generates random numbers by measuring physical phenomena, such as thermal noise or
- radioactive decay. These sources of randomness are considered to be more reliable and unpredictable than
- software-based sources.
-
- One such hardware RNG is the Trusted Platform Module (TPM), which is a dedicated hardware chip that is used
- for cryptographic operations and secure boot. The TPM contains a built-in hardware RNG that generates
- high-quality entropy, which can be used to seed the pool of entropy in the operating system.
-
- Alternatively, software-based solutions such as `Haveged` can be used to generate additional entropy by
- exploiting sources of randomness in the system, such as CPU utilization and network traffic. These solutions
- can help to mitigate the risk of entropy depletion, but they may not be as reliable as hardware-based solutions.
- method_description: ""
- supported_platforms:
- include: ["linux"]
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: lowest_entropy
- link: https://github.com/netdata/netdata/blob/master/health/health.d/entropy.conf
- metric: system.entropy
- info: minimum number of bits of entropy available for the kernel’s random number generator
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.entropy
- description: Available Entropy
- unit: "entropy"
- chart_type: line
- dimensions:
- - name: entropy
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/uptime
- monitored_instance:
- name: System Uptime
- link: ""
- categories:
- - data-collection.linux-systems.system-metrics
- icon_filename: "linuxserver.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - uptime
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- The amount of time the system has been up (running).
-
- Uptime is a critical aspect of overall system performance:
-
- - **Availability**: Uptime monitoring can show whether a server is consistently available or experiences frequent downtimes.
- - **Performance Monitoring**: While server uptime alone doesn't provide detailed performance data, analyzing the duration and frequency of downtimes can help identify patterns or trends.
- - **Proactive problem detection**: If server uptime monitoring reveals unexpected downtimes or a decreasing uptime trend, it can serve as an early warning sign of potential problems.
- - **Root cause analysis**: When investigating server downtime, the uptime metric alone may not provide enough information to pinpoint the exact cause.
- - **Load balancing**: Uptime data can indirectly indicate load balancing issues if certain servers have significantly lower uptimes than others.
- - **Optimize maintenance efforts**: Servers with consistently low uptimes or frequent downtimes may require more attention.
- - **Compliance requirements**: Server uptime data can be used to demonstrate compliance with regulatory requirements or SLAs that mandate a minimum level of server availability.
- method_description: ""
- supported_platforms:
- include: ["linux"]
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.uptime
- description: System Uptime
- unit: "seconds"
- chart_type: line
- dimensions:
- - name: uptime
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/vmstat
- monitored_instance:
- name: Memory Statistics
- link: ""
- categories:
- - data-collection.linux-systems.memory-metrics
- icon_filename: "linuxserver.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - swap
- - page faults
- - oom
- - numa
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- Linux Virtual memory subsystem.
-
- Information about memory management, indicating how effectively the kernel allocates and frees
- memory resources in response to system demands.
-
- Monitors page faults, which occur when a process requests a portion of its memory that isn't
- immediately available. Monitoring these events can help diagnose inefficiencies in memory management and
- provide insights into application behavior.
-
- Tracks swapping activity — a vital aspect of memory management where the kernel moves data from RAM to
- swap space, and vice versa, based on memory demand and usage. It also monitors the utilization of zswap,
- a compressed cache for swap pages, and provides insights into its usage and performance implications.
-
- In the context of virtualized environments, it tracks the ballooning mechanism which is used to balance
- memory resources between host and guest systems.
-
- For systems using NUMA architecture, it provides insights into the local and remote memory accesses, which
- can impact the performance based on the memory access times.
-
- The collector also watches for 'Out of Memory' kills, a drastic measure taken by the system when it runs out
- of memory resources.
- method_description: ""
- supported_platforms:
- include: ["linux"]
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 30min_ram_swapped_out
- link: https://github.com/netdata/netdata/blob/master/health/health.d/swap.conf
- metric: mem.swapio
- info: percentage of the system RAM swapped in the last 30 minutes
- os: "linux freebsd"
- - name: oom_kill
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf
- metric: mem.oom_kill
- info: number of out of memory kills in the last 30 minutes
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: mem.swapio
- description: Swap I/O
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: in
- - name: out
- - name: system.pgpgio
- description: Memory Paged from/to disk
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: in
- - name: out
- - name: system.pgfaults
- description: Memory Page Faults
- unit: "faults/s"
- chart_type: line
- dimensions:
- - name: minor
- - name: major
- - name: mem.balloon
- description: Memory Ballooning Operations
- unit: "KiB/s"
- chart_type: line
- dimensions:
- - name: inflate
- - name: deflate
- - name: migrate
- - name: mem.zswapio
- description: ZSwap I/O
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: in
- - name: out
- - name: mem.ksm_cow
- description: KSM Copy On Write Operations
- unit: "KiB/s"
- chart_type: line
- dimensions:
- - name: swapin
- - name: write
- - name: mem.thp_faults
- description: Transparent Huge Page Fault Allocations
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: alloc
- - name: fallback
- - name: fallback_charge
- - name: mem.thp_file
- description: Transparent Huge Page File Allocations
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: alloc
- - name: fallback
- - name: mapped
- - name: fallback_charge
- - name: mem.thp_zero
- description: Transparent Huge Zero Page Allocations
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: alloc
- - name: failed
- - name: mem.thp_collapse
- description: Transparent Huge Pages Collapsed by khugepaged
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: alloc
- - name: failed
- - name: mem.thp_split
- description: Transparent Huge Page Splits
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: split
- - name: failed
- - name: split_pmd
- - name: split_deferred
- - name: mem.thp_swapout
- description: Transparent Huge Pages Swap Out
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: swapout
- - name: fallback
- - name: mem.thp_compact
- description: Transparent Huge Pages Compaction
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: success
- - name: fail
- - name: stall
- - name: mem.oom_kill
- description: Out of Memory Kills
- unit: "kills/s"
- chart_type: line
- dimensions:
- - name: kills
- - name: mem.numa
- description: NUMA events
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: local
- - name: foreign
- - name: interleave
- - name: other
- - name: pte_updates
- - name: huge_pte_updates
- - name: hint_faults
- - name: hint_faults_local
- - name: pages_migrated
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/interrupts
- monitored_instance:
- name: Interrupts
- link: ""
- categories:
- - data-collection.linux-systems.cpu-metrics
- icon_filename: "linuxserver.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - interrupts
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- Monitors `/proc/interrupts`, a file organized by CPU and then by the type of interrupt.
- The numbers reported are the counts of the interrupts that have occurred of each type.
-
- An interrupt is a signal to the processor emitted by hardware or software indicating an event that needs
- immediate attention. The processor then interrupts its current activities and executes the interrupt handler
- to deal with the event. This is part of the way a computer multitasks and handles concurrent processing.
-
- The types of interrupts include:
-
- - **I/O interrupts**: These are caused by I/O devices like the keyboard, mouse, printer, etc. For example, when
- you type something on the keyboard, an interrupt is triggered so the processor can handle the new input.
-
- - **Timer interrupts**: These are generated at regular intervals by the system's timer circuit. It's primarily
- used to switch the CPU among different tasks.
-
- - **Software interrupts**: These are generated by a program requiring disk I/O operations, or other system resources.
-
- - **Hardware interrupts**: These are caused by hardware conditions such as power failure, overheating, etc.
-
- Monitoring `/proc/interrupts` can be used for:
-
- - **Performance tuning**: If an interrupt is happening very frequently, it could be a sign that a device is not
- configured correctly, or there is a software bug causing unnecessary interrupts. This could lead to system
- performance degradation.
-
- - **System troubleshooting**: If you're seeing a lot of unexpected interrupts, it could be a sign of a hardware problem.
-
- - **Understanding system behavior**: More generally, keeping an eye on what interrupts are occurring can help you
- understand what your system is doing. It can provide insights into the system's interaction with hardware,
- drivers, and other parts of the kernel.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.interrupts
- description: System interrupts
- unit: "interrupts/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per device
- - name: cpu core
- description: ""
- labels:
- - name: cpu
- description: TBD
- metrics:
- - name: cpu.interrupts
- description: CPU interrupts
- unit: "interrupts/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per device
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/loadavg
- monitored_instance:
- name: System Load Average
- link: ""
- categories:
- - data-collection.linux-systems.system-metrics
- icon_filename: "linuxserver.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - load
- - load average
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- The `/proc/loadavg` file provides information about the system load average.
-
- The load average is a measure of the amount of computational work that a system performs. It is a
- representation of the average system load over a period of time.
-
- This file contains three numbers representing the system load averages for the last 1, 5, and 15 minutes,
- respectively. It also includes the currently running processes and the total number of processes.
-
- Monitoring the load average can be used for:
-
- - **System performance**: If the load average is too high, it may indicate that your system is overloaded.
- On a system with a single CPU, if the load average is 1, it means the single CPU is fully utilized. If the
- load averages are consistently higher than the number of CPUs/cores, it may indicate that your system is
- overloaded and tasks are waiting for CPU time.
-
- - **Troubleshooting**: If the load average is unexpectedly high, it can be a sign of a problem. This could be
- due to a runaway process, a software bug, or a hardware issue.
-
- - **Capacity planning**: By monitoring the load average over time, you can understand the trends in your
- system's workload. This can help with capacity planning and scaling decisions.
-
- Remember that load average not only considers CPU usage, but also includes processes waiting for disk I/O.
- Therefore, high load averages could be due to I/O contention as well as CPU contention.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: load_cpu_number
- link: https://github.com/netdata/netdata/blob/master/health/health.d/load.conf
- metric: system.load
- info: number of active CPU cores in the system
- os: "linux"
- - name: load_average_15
- link: https://github.com/netdata/netdata/blob/master/health/health.d/load.conf
- metric: system.load
- info: system fifteen-minute load average
- os: "linux"
- - name: load_average_5
- link: https://github.com/netdata/netdata/blob/master/health/health.d/load.conf
- metric: system.load
- info: system five-minute load average
- os: "linux"
- - name: load_average_1
- link: https://github.com/netdata/netdata/blob/master/health/health.d/load.conf
- metric: system.load
- info: system one-minute load average
- os: "linux"
- - name: active_processes
- link: https://github.com/netdata/netdata/blob/master/health/health.d/processes.conf
- metric: system.active_processes
- info: system process IDs (PID) space utilization
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.load
- description: System Load Average
- unit: "load"
- chart_type: line
- dimensions:
- - name: load1
- - name: load5
- - name: load15
- - name: system.active_processes
- description: System Active Processes
- unit: "processes"
- chart_type: line
- dimensions:
- - name: active
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/pressure
- monitored_instance:
- name: Pressure Stall Information
- link: ""
- categories:
- - data-collection.linux-systems.pressure-metrics
- icon_filename: "linuxserver.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - pressure
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- Introduced in Linux kernel 4.20, `/proc/pressure` provides information about system pressure stall information
- (PSI). PSI is a feature that allows the system to track the amount of time the system is stalled due to
- resource contention, such as CPU, memory, or I/O.
-
- The collectors monitored 3 separate files for CPU, memory, and I/O:
-
- - **cpu**: Tracks the amount of time tasks are stalled due to CPU contention.
- - **memory**: Tracks the amount of time tasks are stalled due to memory contention.
- - **io**: Tracks the amount of time tasks are stalled due to I/O contention.
- - **irq**: Tracks the amount of time tasks are stalled due to IRQ contention.
-
- Each of them provides metrics for stall time over the last 10 seconds, 1 minute, 5 minutes, and 15 minutes.
-
- Monitoring the /proc/pressure files can provide important insights into system performance and capacity planning:
-
- - **Identifying resource contention**: If these metrics are consistently high, it indicates that tasks are
- frequently being stalled due to lack of resources, which can significantly degrade system performance.
-
- - **Troubleshooting performance issues**: If a system is experiencing performance issues, these metrics can
- help identify whether resource contention is the cause.
-
- - **Capacity planning**: By monitoring these metrics over time, you can understand trends in resource
- utilization and make informed decisions about when to add more resources to your system.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.cpu_some_pressure
- description: CPU some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: system.cpu_some_pressure_stall_time
- description: CPU some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: system.cpu_full_pressure
- description: CPU full pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: system.cpu_full_pressure_stall_time
- description: CPU full pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: system.memory_some_pressure
- description: Memory some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: system.memory_some_pressure_stall_time
- description: Memory some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: system.memory_full_pressure
- description: Memory full pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: system.memory_full_pressure_stall_time
- description: Memory full pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: system.io_some_pressure
- description: I/O some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: system.io_some_pressure_stall_time
- description: I/O some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - name: system.io_full_pressure
- description: I/O some pressure
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: some10
- - name: some60
- - name: some300
- - name: system.io_full_pressure_stall_time
- description: I/O some pressure stall time
- unit: "ms"
- chart_type: line
- dimensions:
- - name: time
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/softirqs
- monitored_instance:
- name: SoftIRQ statistics
- link: ""
- categories:
- - data-collection.linux-systems.cpu-metrics
- icon_filename: "linuxserver.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - softirqs
- - interrupts
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- In the Linux kernel, handling of hardware interrupts is split into two halves: the top half and the bottom half.
- The top half is the routine that responds immediately to an interrupt, while the bottom half is deferred to be processed later.
-
- Softirqs are a mechanism in the Linux kernel used to handle the bottom halves of interrupts, which can be
- deferred and processed later in a context where it's safe to enable interrupts.
-
- The actual work of handling the interrupt is offloaded to a softirq and executed later when the system
- decides it's a good time to process them. This helps to keep the system responsive by not blocking the top
- half for too long, which could lead to missed interrupts.
-
- Monitoring `/proc/softirqs` is useful for:
-
- - **Performance tuning**: A high rate of softirqs could indicate a performance issue. For instance, a high
- rate of network softirqs (`NET_RX` and `NET_TX`) could indicate a network performance issue.
-
- - **Troubleshooting**: If a system is behaving unexpectedly, checking the softirqs could provide clues about
- what is going on. For example, a sudden increase in block device softirqs (BLOCK) might indicate a problem
- with a disk.
-
- - **Understanding system behavior**: Knowing what types of softirqs are happening can help you understand what
- your system is doing, particularly in terms of how it's interacting with hardware and how it's handling
- interrupts.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.softirqs
- description: System softirqs
- unit: "softirqs/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per softirq
- - name: cpu core
- description: ""
- labels:
- - name: cpu
- description: TBD
- metrics:
- - name: cpu.softirqs
- description: CPU softirqs
- unit: "softirqs/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per softirq
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/softnet_stat
- monitored_instance:
- name: Softnet Statistics
- link: ""
- categories:
- - data-collection.linux-systems.network-metrics
- icon_filename: "linuxserver.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - softnet
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- `/proc/net/softnet_stat` provides statistics that relate to the handling of network packets by softirq.
-
- It provides information about:
-
- - Total number of processed packets (`processed`).
- - Times ksoftirq ran out of quota (`dropped`).
- - Times net_rx_action was rescheduled.
- - Number of times processed all lists before quota.
- - Number of times did not process all lists due to quota.
- - Number of times net_rx_action was rescheduled for GRO (Generic Receive Offload) cells.
- - Number of times GRO cells were processed.
-
- Monitoring the /proc/net/softnet_stat file can be useful for:
-
- - **Network performance monitoring**: By tracking the total number of processed packets and how many packets
- were dropped, you can gain insights into your system's network performance.
-
- - **Troubleshooting**: If you're experiencing network-related issues, this collector can provide valuable clues.
- For instance, a high number of dropped packets may indicate a network problem.
-
- - **Capacity planning**: If your system is consistently processing near its maximum capacity of network
- packets, it might be time to consider upgrading your network infrastructure.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 1min_netdev_backlog_exceeded
- link: https://github.com/netdata/netdata/blob/master/health/health.d/softnet.conf
- metric: system.softnet_stat
- info: average number of dropped packets in the last minute due to exceeded net.core.netdev_max_backlog
- os: "linux"
- - name: 1min_netdev_budget_ran_outs
- link: https://github.com/netdata/netdata/blob/master/health/health.d/softnet.conf
- metric: system.softnet_stat
- info:
- average number of times ksoftirq ran out of sysctl net.core.netdev_budget or net.core.netdev_budget_usecs with work remaining over the last
- minute (this can be a cause for dropped packets)
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.softnet_stat
- description: System softnet_stat
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: processed
- - name: dropped
- - name: squeezed
- - name: received_rps
- - name: flow_limit_count
- - name: cpu core
- description: ""
- labels: []
- metrics:
- - name: cpu.softnet_stat
- description: CPU softnet_stat
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: processed
- - name: dropped
- - name: squeezed
- - name: received_rps
- - name: flow_limit_count
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/meminfo
- monitored_instance:
- name: Memory Usage
- link: ""
- categories:
- - data-collection.linux-systems.memory-metrics
- icon_filename: "linuxserver.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - memory
- - ram
- - available
- - committed
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- `/proc/meminfo` provides detailed information about the system's current memory usage. It includes information
- about different types of memory, RAM, Swap, ZSwap, HugePages, Transparent HugePages (THP), Kernel memory,
- SLAB memory, memory mappings, and more.
-
- Monitoring /proc/meminfo can be useful for:
-
- - **Performance Tuning**: Understanding your system's memory usage can help you make decisions about system
- tuning and optimization. For example, if your system is frequently low on free memory, it might benefit
- from more RAM.
-
- - **Troubleshooting**: If your system is experiencing problems, `/proc/meminfo` can provide clues about
- whether memory usage is a factor. For example, if your system is slow and cached swap is high, it could
- mean that your system is swapping out a lot of memory to disk, which can degrade performance.
-
- - **Capacity Planning**: By monitoring memory usage over time, you can understand trends and make informed
- decisions about future capacity needs.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: ram_in_use
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf
- metric: system.ram
- info: system memory utilization
- os: "linux"
- - name: ram_available
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf
- metric: mem.available
- info: percentage of estimated amount of RAM available for userspace processes, without causing swapping
- os: "linux"
- - name: used_swap
- link: https://github.com/netdata/netdata/blob/master/health/health.d/swap.conf
- metric: mem.swap
- info: swap memory utilization
- os: "linux freebsd"
- - name: 1hour_memory_hw_corrupted
- link: https://github.com/netdata/netdata/blob/master/health/health.d/memory.conf
- metric: mem.hwcorrupt
- info: amount of memory corrupted due to a hardware failure
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.ram
- description: System RAM
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: free
- - name: used
- - name: cached
- - name: buffers
- - name: mem.available
- description: Available RAM for applications
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: avail
- - name: mem.swap
- description: System Swap
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: free
- - name: used
- - name: mem.swap_cached
- description: Swap Memory Cached in RAM
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: cached
- - name: mem.zswap
- description: Zswap Usage
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: in-ram
- - name: on-disk
- - name: mem.hwcorrupt
- description: Corrupted Memory detected by ECC
- unit: "MiB"
- chart_type: line
- dimensions:
- - name: HardwareCorrupted
- - name: mem.commited
- description: Committed (Allocated) Memory
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: Commited_AS
- - name: mem.writeback
- description: Writeback Memory
- unit: "MiB"
- chart_type: line
- dimensions:
- - name: Dirty
- - name: Writeback
- - name: FuseWriteback
- - name: NfsWriteback
- - name: Bounce
- - name: mem.kernel
- description: Memory Used by Kernel
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: Slab
- - name: KernelStack
- - name: PageTables
- - name: VmallocUsed
- - name: Percpu
- - name: mem.slab
- description: Reclaimable Kernel Memory
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: reclaimable
- - name: unreclaimable
- - name: mem.hugepages
- description: Dedicated HugePages Memory
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: free
- - name: used
- - name: surplus
- - name: reserved
- - name: mem.thp
- description: Transparent HugePages Memory
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: anonymous
- - name: shmem
- - name: mem.thp_details
- description: Details of Transparent HugePages Usage
- unit: "MiB"
- chart_type: line
- dimensions:
- - name: ShmemPmdMapped
- - name: FileHugePages
- - name: FilePmdMapped
- - name: mem.reclaiming
- description: Memory Reclaiming
- unit: "MiB"
- chart_type: line
- dimensions:
- - name: Active
- - name: Inactive
- - name: Active(anon)
- - name: Inactive(anon)
- - name: Active(file)
- - name: Inactive(file)
- - name: Unevictable
- - name: Mlocked
- - name: mem.high_low
- description: High and Low Used and Free Memory Areas
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: high_used
- - name: low_used
- - name: high_free
- - name: low_free
- - name: mem.cma
- description: Contiguous Memory Allocator (CMA) Memory
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: used
- - name: free
- - name: mem.directmaps
- description: Direct Memory Mappings
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: 4k
- - name: 2m
- - name: 4m
- - name: 1g
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/pagetypeinfo
- monitored_instance:
- name: Page types
- link: ""
- categories:
- - data-collection.linux-systems.memory-metrics
- icon_filename: "microchip.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - memory page types
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration provides metrics about the system's memory page types"
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: mem.pagetype_global
- description: System orders available
- unit: "B"
- chart_type: stacked
- dimensions:
- - name: a dimension per pagesize
- - name: node, zone, type
- description: ""
- labels:
- - name: node_id
- description: TBD
- - name: node_zone
- description: TBD
- - name: node_type
- description: TBD
- metrics:
- - name: mem.pagetype
- description: pagetype_Node{node}_{zone}_{type}
- unit: "B"
- chart_type: stacked
- dimensions:
- - name: a dimension per pagesize
- - meta:
- plugin_name: proc.plugin
- module_name: /sys/devices/system/edac/mc
- monitored_instance:
- name: Memory modules (DIMMs)
- link: ""
- categories:
- - data-collection.linux-systems.memory-metrics
- icon_filename: "microchip.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - edac
- - ecc
- - dimm
- - ram
- - hardware
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- The Error Detection and Correction (EDAC) subsystem is detecting and reporting errors in the system's memory,
- primarily ECC (Error-Correcting Code) memory errors.
-
- The collector provides data for:
-
- - Per memory controller (MC): correctable and uncorrectable errors. These can be of 2 kinds:
- - errors related to a DIMM
- - errors that cannot be associated with a DIMM
-
- - Per memory DIMM: correctable and uncorrectable errors. There are 2 kinds:
- - memory controllers that can identify the physical DIMMS and report errors directly for them,
- - memory controllers that report errors for memory address ranges that can be linked to dimms.
- In this case the DIMMS reported may be more than the physical DIMMS installed.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: ecc_memory_mc_noinfo_correctable
- metric: mem.edac_mc
- info: memory controller ${label:controller} ECC correctable errors (unknown DIMM slot) in the last 10 minutes
- link: https://github.com/netdata/netdata/blob/master/health/health.d/memory.conf
- - name: ecc_memory_mc_noinfo_uncorrectable
- metric: mem.edac_mc
- info: memory controller ${label:controller} ECC uncorrectable errors (unknown DIMM slot) in the last 10 minutes
- link: https://github.com/netdata/netdata/blob/master/health/health.d/memory.conf
- - name: ecc_memory_dimm_correctable
- metric: mem.edac_mc_dimm
- info: DIMM ${label:dimm} controller ${label:controller} (location ${label:dimm_location}) ECC correctable errors in the last 10 minutes
- link: https://github.com/netdata/netdata/blob/master/health/health.d/memory.conf
- - name: ecc_memory_dimm_uncorrectable
- metric: mem.edac_mc_dimm
- info: DIMM ${label:dimm} controller ${label:controller} (location ${label:dimm_location}) ECC uncorrectable errors in the last 10 minutes
- link: https://github.com/netdata/netdata/blob/master/health/health.d/memory.conf
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: memory controller
- description: These metrics refer to the memory controller.
- labels:
- - name: controller
- description: "[mcX](https://www.kernel.org/doc/html/v5.0/admin-guide/ras.html#mcx-directories) directory name of this memory controller."
- - name: mc_name
- description: Memory controller type.
- - name: size_mb
- description: The amount of memory in megabytes that this memory controller manages.
- - name: max_location
- description: Last available memory slot in this memory controller.
- metrics:
- - name: mem.edac_mc
- description: Memory Controller (MC) Error Detection And Correction (EDAC) Errors
- unit: errors/s
- chart_type: line
- dimensions:
- - name: correctable
- - name: uncorrectable
- - name: correctable_noinfo
- - name: uncorrectable_noinfo
- - name: memory module
- description: These metrics refer to the memory module (or rank, [depends on the memory controller](https://www.kernel.org/doc/html/v5.0/admin-guide/ras.html#f5)).
- labels:
- - name: controller
- description: "[mcX](https://www.kernel.org/doc/html/v5.0/admin-guide/ras.html#mcx-directories) directory name of this memory controller."
- - name: dimm
- description: "[dimmX or rankX](https://www.kernel.org/doc/html/v5.0/admin-guide/ras.html#dimmx-or-rankx-directories) directory name of this memory module."
- - name: dimm_dev_type
- description: Type of DRAM device used in this memory module. For example, x1, x2, x4, x8.
- - name: dimm_edac_mode
- description: Used type of error detection and correction. For example, S4ECD4ED would mean a Chipkill with x4 DRAM.
- - name: dimm_label
- description: Label assigned to this memory module.
- - name: dimm_location
- description: Location of the memory module.
- - name: dimm_mem_type
- description: Type of the memory module.
- - name: size
- description: The amount of memory in megabytes that this memory module manages.
- metrics:
- - name: mem.edac_mc
- description: DIMM Error Detection And Correction (EDAC) Errors
- unit: errors/s
- chart_type: line
- dimensions:
- - name: correctable
- - name: uncorrectable
- - meta:
- plugin_name: proc.plugin
- module_name: /sys/devices/system/node
- monitored_instance:
- name: Non-Uniform Memory Access
- link: ""
- categories:
- - data-collection.linux-systems.memory-metrics
- icon_filename: "linuxserver.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - numa
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- Information about NUMA (Non-Uniform Memory Access) nodes on the system.
-
- NUMA is a method of configuring a cluster of microprocessor in a multiprocessing system so that they can
- share memory locally, improving performance and the ability of the system to be expanded. NUMA is used in a
- symmetric multiprocessing (SMP) system.
-
- In a NUMA system, processors, memory, and I/O devices are grouped together into cells, also known as nodes.
- Each node has its own memory and set of I/O devices, and one or more processors. While a processor can access
- memory in any of the nodes, it does so faster when accessing memory within its own node.
-
- The collector provides statistics on memory allocations for processes running on the NUMA nodes, revealing the
- efficiency of memory allocations in multi-node systems.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: numa node
- description: ""
- labels:
- - name: numa_node
- description: TBD
- metrics:
- - name: mem.numa_nodes
- description: NUMA events
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: hit
- - name: miss
- - name: local
- - name: foreign
- - name: interleave
- - name: other
- - meta:
- plugin_name: proc.plugin
- module_name: /sys/kernel/mm/ksm
- monitored_instance:
- name: Kernel Same-Page Merging
- link: ""
- categories:
- - data-collection.linux-systems.memory-metrics
- icon_filename: "microchip.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - ksm
- - samepage
- - merging
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- Kernel Samepage Merging (KSM) is a memory-saving feature in Linux that enables the kernel to examine the
- memory of different processes and identify identical pages. It then merges these identical pages into a
- single page that the processes share. This is particularly useful for virtualization, where multiple virtual
- machines might be running the same operating system or applications and have many identical pages.
-
- The collector provides information about the operation and effectiveness of KSM on your system.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: mem.ksm
- description: Kernel Same Page Merging
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: shared
- - name: unshared
- - name: sharing
- - name: volatile
- - name: mem.ksm_savings
- description: Kernel Same Page Merging Savings
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: savings
- - name: offered
- - name: mem.ksm_ratios
- description: Kernel Same Page Merging Effectiveness
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: savings
- - meta:
- plugin_name: proc.plugin
- module_name: /sys/block/zram
- monitored_instance:
- name: ZRAM
- link: ""
- categories:
- - data-collection.linux-systems.memory-metrics
- icon_filename: "microchip.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - zram
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- zRAM, or compressed RAM, is a block device that uses a portion of your system's RAM as a block device.
- The data written to this block device is compressed and stored in memory.
-
- The collectors provides information about the operation and the effectiveness of zRAM on your system.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: zram device
- description: ""
- labels:
- - name: device
- description: TBD
- metrics:
- - name: mem.zram_usage
- description: ZRAM Memory Usage
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: compressed
- - name: metadata
- - name: mem.zram_savings
- description: ZRAM Memory Savings
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: savings
- - name: original
- - name: mem.zram_ratio
- description: ZRAM Compression Ratio (original to compressed)
- unit: "ratio"
- chart_type: line
- dimensions:
- - name: ratio
- - name: mem.zram_efficiency
- description: ZRAM Efficiency
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: percent
- - meta:
- plugin_name: proc.plugin
- module_name: ipc
- monitored_instance:
- name: Inter Process Communication
- link: ""
- categories:
- - data-collection.linux-systems.ipc-metrics
- icon_filename: "network-wired.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - ipc
- - semaphores
- - shared memory
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- IPC stands for Inter-Process Communication. It is a mechanism which allows processes to communicate with each
- other and synchronize their actions.
-
- This collector exposes information about:
-
- - Message Queues: This allows messages to be exchanged between processes. It's a more flexible method that
- allows messages to be placed onto a queue and read at a later time.
-
- - Shared Memory: This method allows for the fastest form of IPC because processes can exchange data by
- reading/writing into shared memory segments.
-
- - Semaphores: They are used to synchronize the operations performed by independent processes. So, if multiple
- processes are trying to access a single shared resource, semaphores can ensure that only one process
- accesses the resource at a given time.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: semaphores_used
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ipc.conf
- metric: system.ipc_semaphores
- info: IPC semaphore utilization
- os: "linux"
- - name: semaphore_arrays_used
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ipc.conf
- metric: system.ipc_semaphore_arrays
- info: IPC semaphore arrays utilization
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.ipc_semaphores
- description: IPC Semaphores
- unit: "semaphores"
- chart_type: area
- dimensions:
- - name: semaphores
- - name: system.ipc_semaphore_arrays
- description: IPC Semaphore Arrays
- unit: "arrays"
- chart_type: area
- dimensions:
- - name: arrays
- - name: system.message_queue_message
- description: IPC Message Queue Number of Messages
- unit: "messages"
- chart_type: stacked
- dimensions:
- - name: a dimension per queue
- - name: system.message_queue_bytes
- description: IPC Message Queue Used Bytes
- unit: "bytes"
- chart_type: stacked
- dimensions:
- - name: a dimension per queue
- - name: system.shared_memory_segments
- description: IPC Shared Memory Number of Segments
- unit: "segments"
- chart_type: stacked
- dimensions:
- - name: segments
- - name: system.shared_memory_bytes
- description: IPC Shared Memory Used Bytes
- unit: "bytes"
- chart_type: stacked
- dimensions:
- - name: bytes
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/diskstats
- monitored_instance:
- name: Disk Statistics
- link: ""
- categories:
- - data-collection.linux-systems.disk-metrics
- icon_filename: "hard-drive.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - disk
- - disks
- - io
- - bcache
- - block devices
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- Detailed statistics for each of your system's disk devices and partitions.
- The data is reported by the kernel and can be used to monitor disk activity on a Linux system.
-
- Get valuable insight into how your disks are performing and where potential bottlenecks might be.
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 10min_disk_backlog
- link: https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf
- metric: disk.backlog
- info: average backlog size of the ${label:device} disk over the last 10 minutes
- os: "linux"
- - name: 10min_disk_utilization
- link: https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf
- metric: disk.util
- info: average percentage of time ${label:device} disk was busy over the last 10 minutes
- os: "linux freebsd"
- - name: bcache_cache_dirty
- link: https://github.com/netdata/netdata/blob/master/health/health.d/bcache.conf
- metric: disk.bcache_cache_alloc
- info: percentage of cache space used for dirty data and metadata (this usually means your SSD cache is too small)
- - name: bcache_cache_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/bcache.conf
- metric: disk.bcache_cache_read_races
- info:
- number of times data was read from the cache, the bucket was reused and invalidated in the last 10 minutes (when this occurs the data is
- reread from the backing device)
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.io
- description: Disk I/O
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: in
- - name: out
- - name: disk
- description: ""
- labels:
- - name: device
- description: TBD
- - name: mount_point
- description: TBD
- - name: device_type
- description: TBD
- metrics:
- - name: disk.io
- description: Disk I/O Bandwidth
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: reads
- - name: writes
- - name: disk_ext.io
- description: Amount of Discarded Data
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: discards
- - name: disk.ops
- description: Disk Completed I/O Operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: reads
- - name: writes
- - name: disk_ext.ops
- description: Disk Completed Extended I/O Operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: discards
- - name: flushes
- - name: disk.qops
- description: Disk Current I/O Operations
- unit: "operations"
- chart_type: line
- dimensions:
- - name: operations
- - name: disk.backlog
- description: Disk Backlog
- unit: "milliseconds"
- chart_type: area
- dimensions:
- - name: backlog
- - name: disk.busy
- description: Disk Busy Time
- unit: "milliseconds"
- chart_type: area
- dimensions:
- - name: busy
- - name: disk.util
- description: Disk Utilization Time
- unit: "% of time working"
- chart_type: area
- dimensions:
- - name: utilization
- - name: disk.mops
- description: Disk Merged Operations
- unit: "merged operations/s"
- chart_type: line
- dimensions:
- - name: reads
- - name: writes
- - name: disk_ext.mops
- description: Disk Merged Discard Operations
- unit: "merged operations/s"
- chart_type: line
- dimensions:
- - name: discards
- - name: disk.iotime
- description: Disk Total I/O Time
- unit: "milliseconds/s"
- chart_type: line
- dimensions:
- - name: reads
- - name: writes
- - name: disk_ext.iotime
- description: Disk Total I/O Time for Extended Operations
- unit: "milliseconds/s"
- chart_type: line
- dimensions:
- - name: discards
- - name: flushes
- - name: disk.await
- description: Average Completed I/O Operation Time
- unit: "milliseconds/operation"
- chart_type: line
- dimensions:
- - name: reads
- - name: writes
- - name: disk_ext.await
- description: Average Completed Extended I/O Operation Time
- unit: "milliseconds/operation"
- chart_type: line
- dimensions:
- - name: discards
- - name: flushes
- - name: disk.avgsz
- description: Average Completed I/O Operation Bandwidth
- unit: "KiB/operation"
- chart_type: area
- dimensions:
- - name: reads
- - name: writes
- - name: disk_ext.avgsz
- description: Average Amount of Discarded Data
- unit: "KiB/operation"
- chart_type: area
- dimensions:
- - name: discards
- - name: disk.svctm
- description: Average Service Time
- unit: "milliseconds/operation"
- chart_type: line
- dimensions:
- - name: svctm
- - name: disk.bcache_cache_alloc
- description: BCache Cache Allocations
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: ununsed
- - name: dirty
- - name: clean
- - name: metadata
- - name: undefined
- - name: disk.bcache_hit_ratio
- description: BCache Cache Hit Ratio
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: 5min
- - name: 1hour
- - name: 1day
- - name: ever
- - name: disk.bcache_rates
- description: BCache Rates
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: congested
- - name: writeback
- - name: disk.bcache_size
- description: BCache Cache Sizes
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: dirty
- - name: disk.bcache_usage
- description: BCache Cache Usage
- unit: "percentage"
- chart_type: area
- dimensions:
- - name: avail
- - name: disk.bcache_cache_read_races
- description: BCache Cache Read Races
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: races
- - name: errors
- - name: disk.bcache
- description: BCache Cache I/O Operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: hits
- - name: misses
- - name: collisions
- - name: readaheads
- - name: disk.bcache_bypass
- description: BCache Cache Bypass I/O Operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: hits
- - name: misses
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/mdstat
- monitored_instance:
- name: MD RAID
- link: ""
- categories:
- - data-collection.linux-systems.disk-metrics
- icon_filename: "hard-drive.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - raid
- - mdadm
- - mdstat
- - raid
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration monitors the status of MD RAID devices."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: mdstat_last_collected
- link: https://github.com/netdata/netdata/blob/master/health/health.d/mdstat.conf
- metric: md.disks
- info: number of seconds since the last successful data collection
- - name: mdstat_disks
- link: https://github.com/netdata/netdata/blob/master/health/health.d/mdstat.conf
- metric: md.disks
- info:
- number of devices in the down state for the ${label:device} ${label:raid_level} array. Any number > 0 indicates that the array is degraded.
- - name: mdstat_mismatch_cnt
- link: https://github.com/netdata/netdata/blob/master/health/health.d/mdstat.conf
- metric: md.mismatch_cnt
- info: number of unsynchronized blocks for the ${label:device} ${label:raid_level} array
- - name: mdstat_nonredundant_last_collected
- link: https://github.com/netdata/netdata/blob/master/health/health.d/mdstat.conf
- metric: md.nonredundant
- info: number of seconds since the last successful data collection
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: md.health
- description: Faulty Devices In MD
- unit: "failed disks"
- chart_type: line
- dimensions:
- - name: a dimension per md array
- - name: md array
- description: ""
- labels:
- - name: device
- description: TBD
- - name: raid_level
- description: TBD
- metrics:
- - name: md.disks
- description: Disks Stats
- unit: "disks"
- chart_type: stacked
- dimensions:
- - name: inuse
- - name: down
- - name: md.mismatch_cnt
- description: Mismatch Count
- unit: "unsynchronized blocks"
- chart_type: line
- dimensions:
- - name: count
- - name: md.status
- description: Current Status
- unit: "percent"
- chart_type: line
- dimensions:
- - name: check
- - name: resync
- - name: recovery
- - name: reshape
- - name: md.expected_time_until_operation_finish
- description: Approximate Time Until Finish
- unit: "seconds"
- chart_type: line
- dimensions:
- - name: finish_in
- - name: md.operation_speed
- description: Operation Speed
- unit: "KiB/s"
- chart_type: line
- dimensions:
- - name: speed
- - name: md.nonredundant
- description: Nonredundant Array Availability
- unit: "boolean"
- chart_type: line
- dimensions:
- - name: available
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/dev
- monitored_instance:
- name: Network interfaces
- link: ""
- categories:
- - data-collection.linux-systems.network-metrics
- icon_filename: "network-wired.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - network interfaces
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor network interface metrics about bandwidth, state, errors and more."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: interface_speed
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.net
- info: network interface ${label:device} current speed
- os: "*"
- - name: 1m_received_traffic_overflow
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.net
- info: average inbound utilization for the network interface ${label:device} over the last minute
- os: "linux"
- - name: 1m_sent_traffic_overflow
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.net
- info: average outbound utilization for the network interface ${label:device} over the last minute
- os: "linux"
- - name: inbound_packets_dropped_ratio
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.drops
- info: ratio of inbound dropped packets for the network interface ${label:device} over the last 10 minutes
- os: "linux"
- - name: outbound_packets_dropped_ratio
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.drops
- info: ratio of outbound dropped packets for the network interface ${label:device} over the last 10 minutes
- os: "linux"
- - name: wifi_inbound_packets_dropped_ratio
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.drops
- info: ratio of inbound dropped packets for the network interface ${label:device} over the last 10 minutes
- os: "linux"
- - name: wifi_outbound_packets_dropped_ratio
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.drops
- info: ratio of outbound dropped packets for the network interface ${label:device} over the last 10 minutes
- os: "linux"
- - name: 1m_received_packets_rate
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.packets
- info: average number of packets received by the network interface ${label:device} over the last minute
- os: "linux freebsd"
- - name: 10s_received_packets_storm
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.packets
- info: ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute
- os: "linux freebsd"
- - name: 10min_fifo_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/net.conf
- metric: net.fifo
- info: number of FIFO errors for the network interface ${label:device} in the last 10 minutes
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.net
- description: Physical Network Interfaces Aggregated Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: network device
- description: ""
- labels:
- - name: interface_type
- description: TBD
- - name: device
- description: TBD
- metrics:
- - name: net.net
- description: Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: net.speed
- description: Interface Speed
- unit: "kilobits/s"
- chart_type: line
- dimensions:
- - name: speed
- - name: net.duplex
- description: Interface Duplex State
- unit: "state"
- chart_type: line
- dimensions:
- - name: full
- - name: half
- - name: unknown
- - name: net.operstate
- description: Interface Operational State
- unit: "state"
- chart_type: line
- dimensions:
- - name: up
- - name: down
- - name: notpresent
- - name: lowerlayerdown
- - name: testing
- - name: dormant
- - name: unknown
- - name: net.carrier
- description: Interface Physical Link State
- unit: "state"
- chart_type: line
- dimensions:
- - name: up
- - name: down
- - name: net.mtu
- description: Interface MTU
- unit: "octets"
- chart_type: line
- dimensions:
- - name: mtu
- - name: net.packets
- description: Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: multicast
- - name: net.errors
- description: Interface Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: inbound
- - name: outbound
- - name: net.drops
- description: Interface Drops
- unit: "drops/s"
- chart_type: line
- dimensions:
- - name: inbound
- - name: outbound
- - name: net.fifo
- description: Interface FIFO Buffer Errors
- unit: "errors"
- chart_type: line
- dimensions:
- - name: receive
- - name: transmit
- - name: net.compressed
- description: Compressed Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: net.events
- description: Network Interface Events
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: frames
- - name: collisions
- - name: carrier
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/wireless
- monitored_instance:
- name: Wireless network interfaces
- link: ""
- categories:
- - data-collection.linux-systems.network-metrics
- icon_filename: "network-wired.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - wireless devices
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor wireless devices with metrics about status, link quality, signal level, noise level and more."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: wireless device
- description: ""
- labels: []
- metrics:
- - name: wireless.status
- description: Internal status reported by interface.
- unit: "status"
- chart_type: line
- dimensions:
- - name: status
- - name: wireless.link_quality
- description: Overall quality of the link. This is an aggregate value, and depends on the driver and hardware.
- unit: "value"
- chart_type: line
- dimensions:
- - name: link_quality
- - name: wireless.signal_level
- description:
- The signal level is the wireless signal power level received by the wireless client. The closer the value is to 0, the stronger the
- signal.
- unit: "dBm"
- chart_type: line
- dimensions:
- - name: signal_level
- - name: wireless.noise_level
- description:
- The noise level indicates the amount of background noise in your environment. The closer the value to 0, the greater the noise level.
- unit: "dBm"
- chart_type: line
- dimensions:
- - name: noise_level
- - name: wireless.discarded_packets
- description: Packet discarded in the wireless adapter due to wireless specific problems.
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: nwid
- - name: crypt
- - name: frag
- - name: retry
- - name: misc
- - name: wireless.missed_beacons
- description: Number of missed beacons.
- unit: "frames/s"
- chart_type: line
- dimensions:
- - name: missed_beacons
- - meta:
- plugin_name: proc.plugin
- module_name: /sys/class/infiniband
- monitored_instance:
- name: InfiniBand
- link: ""
- categories:
- - data-collection.linux-systems.network-metrics
- icon_filename: "network-wired.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - infiniband
- - rdma
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration monitors InfiniBand network inteface statistics."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: infiniband port
- description: ""
- labels: []
- metrics:
- - name: ib.bytes
- description: Bandwidth usage
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: Received
- - name: Sent
- - name: ib.packets
- description: Packets Statistics
- unit: "packets/s"
- chart_type: area
- dimensions:
- - name: Received
- - name: Sent
- - name: Mcast_rcvd
- - name: Mcast_sent
- - name: Ucast_rcvd
- - name: Ucast_sent
- - name: ib.errors
- description: Error Counters
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: Pkts_malformated
- - name: Pkts_rcvd_discarded
- - name: Pkts_sent_discarded
- - name: Tick_Wait_to_send
- - name: Pkts_missed_resource
- - name: Buffer_overrun
- - name: Link_Downed
- - name: Link_recovered
- - name: Link_integrity_err
- - name: Link_minor_errors
- - name: Pkts_rcvd_with_EBP
- - name: Pkts_rcvd_discarded_by_switch
- - name: Pkts_sent_discarded_by_switch
- - name: ib.hwerrors
- description: Hardware Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: Duplicated_packets
- - name: Pkt_Seq_Num_gap
- - name: Ack_timer_expired
- - name: Drop_missing_buffer
- - name: Drop_out_of_sequence
- - name: NAK_sequence_rcvd
- - name: CQE_err_Req
- - name: CQE_err_Resp
- - name: CQE_Flushed_err_Req
- - name: CQE_Flushed_err_Resp
- - name: Remote_access_err_Req
- - name: Remote_access_err_Resp
- - name: Remote_invalid_req
- - name: Local_length_err_Resp
- - name: RNR_NAK_Packets
- - name: CNP_Pkts_ignored
- - name: RoCE_ICRC_Errors
- - name: ib.hwpackets
- description: Hardware Packets Statistics
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: RoCEv2_Congestion_sent
- - name: RoCEv2_Congestion_rcvd
- - name: IB_Congestion_handled
- - name: ATOMIC_req_rcvd
- - name: Connection_req_rcvd
- - name: Read_req_rcvd
- - name: Write_req_rcvd
- - name: RoCE_retrans_adaptive
- - name: RoCE_retrans_timeout
- - name: RoCE_slow_restart
- - name: RoCE_slow_restart_congestion
- - name: RoCE_slow_restart_count
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/netstat
- monitored_instance:
- name: Network statistics
- link: ""
- categories:
- - data-collection.linux-systems.network-metrics
- icon_filename: "network-wired.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - ip
- - udp
- - udplite
- - icmp
- - netstat
- - snmp
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration provides metrics from the `netstat`, `snmp` and `snmp6` modules."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: 1m_tcp_syn_queue_drops
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_listen.conf
- metric: ip.tcp_syn_queue
- info: average number of SYN requests was dropped due to the full TCP SYN queue over the last minute (SYN cookies were not enabled)
- os: "linux"
- - name: 1m_tcp_syn_queue_cookies
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_listen.conf
- metric: ip.tcp_syn_queue
- info: average number of sent SYN cookies due to the full TCP SYN queue over the last minute
- os: "linux"
- - name: 1m_tcp_accept_queue_overflows
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_listen.conf
- metric: ip.tcp_accept_queue
- info: average number of overflows in the TCP accept queue over the last minute
- os: "linux"
- - name: 1m_tcp_accept_queue_drops
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_listen.conf
- metric: ip.tcp_accept_queue
- info: average number of dropped packets in the TCP accept queue over the last minute
- os: "linux"
- - name: tcp_connections
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_conn.conf
- metric: ip.tcpsock
- info: TCP connections utilization
- os: "linux"
- - name: 1m_ip_tcp_resets_sent
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf
- metric: ip.tcphandshake
- info: average number of sent TCP RESETS over the last minute
- os: "linux"
- - name: 10s_ip_tcp_resets_sent
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf
- metric: ip.tcphandshake
- info:
- average number of sent TCP RESETS over the last 10 seconds. This can indicate a port scan, or that a service running on this host has
- crashed. Netdata will not send a clear notification for this alarm.
- os: "linux"
- - name: 1m_ip_tcp_resets_received
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf
- metric: ip.tcphandshake
- info: average number of received TCP RESETS over the last minute
- os: "linux freebsd"
- - name: 10s_ip_tcp_resets_received
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf
- metric: ip.tcphandshake
- info:
- average number of received TCP RESETS over the last 10 seconds. This can be an indication that a service this host needs has crashed.
- Netdata will not send a clear notification for this alarm.
- os: "linux freebsd"
- - name: 1m_ipv4_udp_receive_buffer_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/udp_errors.conf
- metric: ipv4.udperrors
- info: average number of UDP receive buffer errors over the last minute
- os: "linux freebsd"
- - name: 1m_ipv4_udp_send_buffer_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/udp_errors.conf
- metric: ipv4.udperrors
- info: average number of UDP send buffer errors over the last minute
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: system.ip
- description: IPv4 Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: ip.tcpmemorypressures
- description: TCP Memory Pressures
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: pressures
- - name: ip.tcpconnaborts
- description: TCP Connection Aborts
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: baddata
- - name: userclosed
- - name: nomemory
- - name: timeout
- - name: linger
- - name: failed
- - name: ip.tcpreorders
- description: TCP Reordered Packets by Detection Method
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: timestamp
- - name: sack
- - name: fack
- - name: reno
- - name: ip.tcpofo
- description: TCP Out-Of-Order Queue
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: inqueue
- - name: dropped
- - name: merged
- - name: pruned
- - name: ip.tcpsyncookies
- description: TCP SYN Cookies
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: failed
- - name: ip.tcp_syn_queue
- description: TCP SYN Queue Issues
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: drops
- - name: cookies
- - name: ip.tcp_accept_queue
- description: TCP Accept Queue Issues
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: overflows
- - name: drops
- - name: ip.tcpsock
- description: IPv4 TCP Connections
- unit: "active connections"
- chart_type: line
- dimensions:
- - name: connections
- - name: ip.tcppackets
- description: IPv4 TCP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ip.tcperrors
- description: IPv4 TCP Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InErrs
- - name: InCsumErrors
- - name: RetransSegs
- - name: ip.tcpopens
- description: IPv4 TCP Opens
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: active
- - name: passive
- - name: ip.tcphandshake
- description: IPv4 TCP Handshake Issues
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: EstabResets
- - name: OutRsts
- - name: AttemptFails
- - name: SynRetrans
- - name: ipv4.packets
- description: IPv4 Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: forwarded
- - name: delivered
- - name: ipv4.errors
- description: IPv4 Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InDiscards
- - name: OutDiscards
- - name: InNoRoutes
- - name: OutNoRoutes
- - name: InHdrErrors
- - name: InAddrErrors
- - name: InTruncatedPkts
- - name: InCsumErrors
- - name: ipc4.bcast
- description: IP Broadcast Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.bcastpkts
- description: IP Broadcast Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.mcast
- description: IPv4 Multicast Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.mcastpkts
- description: IP Multicast Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.icmp
- description: IPv4 ICMP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.icmpmsg
- description: IPv4 ICMP Messages
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InEchoReps
- - name: OutEchoReps
- - name: InDestUnreachs
- - name: OutDestUnreachs
- - name: InRedirects
- - name: OutRedirects
- - name: InEchos
- - name: OutEchos
- - name: InRouterAdvert
- - name: OutRouterAdvert
- - name: InRouterSelect
- - name: OutRouterSelect
- - name: InTimeExcds
- - name: OutTimeExcds
- - name: InParmProbs
- - name: OutParmProbs
- - name: InTimestamps
- - name: OutTimestamps
- - name: InTimestampReps
- - name: OutTimestampReps
- - name: ipv4.icmp_errors
- description: IPv4 ICMP Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InErrors
- - name: OutErrors
- - name: InCsumErrors
- - name: ipv4.udppackets
- description: IPv4 UDP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.udperrors
- description: IPv4 UDP Errors
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: RcvbufErrors
- - name: SndbufErrors
- - name: InErrors
- - name: NoPorts
- - name: InCsumErrors
- - name: IgnoredMulti
- - name: ipv4.udplite
- description: IPv4 UDPLite Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv4.udplite_errors
- description: IPv4 UDPLite Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: RcvbufErrors
- - name: SndbufErrors
- - name: InErrors
- - name: NoPorts
- - name: InCsumErrors
- - name: IgnoredMulti
- - name: ipv4.ecnpkts
- description: IP ECN Statistics
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: CEP
- - name: NoECTP
- - name: ECTP0
- - name: ECTP1
- - name: ipv4.fragsin
- description: IPv4 Fragments Reassembly
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: all
- - name: ipv4.fragsout
- description: IPv4 Fragments Sent
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: created
- - name: system.ipv6
- description: IPv6 Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.packets
- description: IPv6 Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: forwarded
- - name: delivers
- - name: ipv6.errors
- description: IPv6 Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InDiscards
- - name: OutDiscards
- - name: InHdrErrors
- - name: InAddrErrors
- - name: InUnknownProtos
- - name: InTooBigErrors
- - name: InTruncatedPkts
- - name: InNoRoutes
- - name: OutNoRoutes
- - name: ipv6.bcast
- description: IPv6 Broadcast Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.mcast
- description: IPv6 Multicast Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.mcastpkts
- description: IPv6 Multicast Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.udppackets
- description: IPv6 UDP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.udperrors
- description: IPv6 UDP Errors
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: RcvbufErrors
- - name: SndbufErrors
- - name: InErrors
- - name: NoPorts
- - name: InCsumErrors
- - name: IgnoredMulti
- - name: ipv6.udplitepackets
- description: IPv6 UDPlite Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.udpliteerrors
- description: IPv6 UDP Lite Errors
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: RcvbufErrors
- - name: SndbufErrors
- - name: InErrors
- - name: NoPorts
- - name: InCsumErrors
- - name: ipv6.icmp
- description: IPv6 ICMP Messages
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.icmpredir
- description: IPv6 ICMP Redirects
- unit: "redirects/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.icmperrors
- description: IPv6 ICMP Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: InErrors
- - name: OutErrors
- - name: InCsumErrors
- - name: InDestUnreachs
- - name: InPktTooBigs
- - name: InTimeExcds
- - name: InParmProblems
- - name: OutDestUnreachs
- - name: OutPktTooBigs
- - name: OutTimeExcds
- - name: OutParmProblems
- - name: ipv6.icmpechos
- description: IPv6 ICMP Echo
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InEchos
- - name: OutEchos
- - name: InEchoReplies
- - name: OutEchoReplies
- - name: ipv6.groupmemb
- description: IPv6 ICMP Group Membership
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InQueries
- - name: OutQueries
- - name: InResponses
- - name: OutResponses
- - name: InReductions
- - name: OutReductions
- - name: ipv6.icmprouter
- description: IPv6 Router Messages
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InSolicits
- - name: OutSolicits
- - name: InAdvertisements
- - name: OutAdvertisements
- - name: ipv6.icmpneighbor
- description: IPv6 Neighbor Messages
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InSolicits
- - name: OutSolicits
- - name: InAdvertisements
- - name: OutAdvertisements
- - name: ipv6.icmpmldv2
- description: IPv6 ICMP MLDv2 Reports
- unit: "reports/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipv6.icmptypes
- description: IPv6 ICMP Types
- unit: "messages/s"
- chart_type: line
- dimensions:
- - name: InType1
- - name: InType128
- - name: InType129
- - name: InType136
- - name: OutType1
- - name: OutType128
- - name: OutType129
- - name: OutType133
- - name: OutType135
- - name: OutType143
- - name: ipv6.ect
- description: IPv6 ECT Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InNoECTPkts
- - name: InECT1Pkts
- - name: InECT0Pkts
- - name: InCEPkts
- - name: ipv6.ect
- description: IPv6 ECT Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: InNoECTPkts
- - name: InECT1Pkts
- - name: InECT0Pkts
- - name: InCEPkts
- - name: ipv6.fragsin
- description: IPv6 Fragments Reassembly
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: timeout
- - name: all
- - name: ipv6.fragsout
- description: IPv6 Fragments Sent
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: ok
- - name: failed
- - name: all
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/sockstat
- monitored_instance:
- name: Socket statistics
- link: ""
- categories:
- - data-collection.linux-systems.network-metrics
- icon_filename: "network-wired.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - sockets
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration provides socket statistics."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: tcp_orphans
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_orphans.conf
- metric: ipv4.sockstat_tcp_sockets
- info: orphan IPv4 TCP sockets utilization
- os: "linux"
- - name: tcp_memory
- link: https://github.com/netdata/netdata/blob/master/health/health.d/tcp_mem.conf
- metric: ipv4.sockstat_tcp_mem
- info: TCP memory utilization
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: ip.sockstat_sockets
- description: Sockets used for all address families
- unit: "sockets"
- chart_type: line
- dimensions:
- - name: used
- - name: ipv4.sockstat_tcp_sockets
- description: IPv4 TCP Sockets
- unit: "sockets"
- chart_type: line
- dimensions:
- - name: alloc
- - name: orphan
- - name: inuse
- - name: timewait
- - name: ipv4.sockstat_tcp_mem
- description: IPv4 TCP Sockets Memory
- unit: "KiB"
- chart_type: area
- dimensions:
- - name: mem
- - name: ipv4.sockstat_udp_sockets
- description: IPv4 UDP Sockets
- unit: "sockets"
- chart_type: line
- dimensions:
- - name: inuse
- - name: ipv4.sockstat_udp_mem
- description: IPv4 UDP Sockets Memory
- unit: "sockets"
- chart_type: line
- dimensions:
- - name: mem
- - name: ipv4.sockstat_udplite_sockets
- description: IPv4 UDPLITE Sockets
- unit: "sockets"
- chart_type: line
- dimensions:
- - name: inuse
- - name: ipv4.sockstat_raw_sockets
- description: IPv4 RAW Sockets
- unit: "sockets"
- chart_type: line
- dimensions:
- - name: inuse
- - name: ipv4.sockstat_frag_sockets
- description: IPv4 FRAG Sockets
- unit: "fragments"
- chart_type: line
- dimensions:
- - name: inuse
- - name: ipv4.sockstat_frag_mem
- description: IPv4 FRAG Sockets Memory
- unit: "KiB"
- chart_type: area
- dimensions:
- - name: mem
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/sockstat6
- monitored_instance:
- name: IPv6 Socket Statistics
- link: ""
- categories:
- - data-collection.linux-systems.network-metrics
- icon_filename: "network-wired.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - ipv6 sockets
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration provides IPv6 socket statistics."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: ipv6.sockstat6_tcp_sockets
- description: IPv6 TCP Sockets
- unit: "sockets"
- chart_type: line
- dimensions:
- - name: inuse
- - name: ipv6.sockstat6_udp_sockets
- description: IPv6 UDP Sockets
- unit: "sockets"
- chart_type: line
- dimensions:
- - name: inuse
- - name: ipv6.sockstat6_udplite_sockets
- description: IPv6 UDPLITE Sockets
- unit: "sockets"
- chart_type: line
- dimensions:
- - name: inuse
- - name: ipv6.sockstat6_raw_sockets
- description: IPv6 RAW Sockets
- unit: "sockets"
- chart_type: line
- dimensions:
- - name: inuse
- - name: ipv6.sockstat6_frag_sockets
- description: IPv6 FRAG Sockets
- unit: "fragments"
- chart_type: line
- dimensions:
- - name: inuse
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/ip_vs_stats
- monitored_instance:
- name: IP Virtual Server
- link: ""
- categories:
- - data-collection.linux-systems.network-metrics
- icon_filename: "network-wired.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - ip virtual server
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration monitors IP Virtual Server statistics"
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: ipvs.sockets
- description: IPVS New Connections
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: connections
- - name: ipvs.packets
- description: IPVS Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: ipvs.net
- description: IPVS Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: received
- - name: sent
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/rpc/nfs
- monitored_instance:
- name: NFS Client
- link: ""
- categories:
- - data-collection.linux-systems.filesystem-metrics.nfs
- icon_filename: "nfs.png"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - nfs client
- - filesystem
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration provides statistics from the Linux kernel's NFS Client."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: nfs.net
- description: NFS Client Network
- unit: "operations/s"
- chart_type: stacked
- dimensions:
- - name: udp
- - name: tcp
- - name: nfs.rpc
- description: NFS Client Remote Procedure Calls Statistics
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: calls
- - name: retransmits
- - name: auth_refresh
- - name: nfs.proc2
- description: NFS v2 Client Remote Procedure Calls
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per proc2 call
- - name: nfs.proc3
- description: NFS v3 Client Remote Procedure Calls
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per proc3 call
- - name: nfs.proc4
- description: NFS v4 Client Remote Procedure Calls
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per proc4 call
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/rpc/nfsd
- monitored_instance:
- name: NFS Server
- link: ""
- categories:
- - data-collection.linux-systems.filesystem-metrics.nfs
- icon_filename: "nfs.png"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - nfs server
- - filesystem
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration provides statistics from the Linux kernel's NFS Server."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: nfsd.readcache
- description: NFS Server Read Cache
- unit: "reads/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: nocache
- - name: nfsd.filehandles
- description: NFS Server File Handles
- unit: "handles/s"
- chart_type: line
- dimensions:
- - name: stale
- - name: nfsd.io
- description: NFS Server I/O
- unit: "kilobytes/s"
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: nfsd.threads
- description: NFS Server Threads
- unit: "threads"
- chart_type: line
- dimensions:
- - name: threads
- - name: nfsd.net
- description: NFS Server Network Statistics
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: udp
- - name: tcp
- - name: nfsd.rpc
- description: NFS Server Remote Procedure Calls Statistics
- unit: "calls/s"
- chart_type: line
- dimensions:
- - name: calls
- - name: bad_format
- - name: bad_auth
- - name: nfsd.proc2
- description: NFS v2 Server Remote Procedure Calls
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per proc2 call
- - name: nfsd.proc3
- description: NFS v3 Server Remote Procedure Calls
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per proc3 call
- - name: nfsd.proc4
- description: NFS v4 Server Remote Procedure Calls
- unit: "calls/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per proc4 call
- - name: nfsd.proc4ops
- description: NFS v4 Server Operations
- unit: "operations/s"
- chart_type: stacked
- dimensions:
- - name: a dimension per proc4 operation
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/sctp/snmp
- monitored_instance:
- name: SCTP Statistics
- link: ""
- categories:
- - data-collection.linux-systems.network-metrics
- icon_filename: "network-wired.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - sctp
- - stream control transmission protocol
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration provides statistics about the Stream Control Transmission Protocol (SCTP)."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: sctp.established
- description: SCTP current total number of established associations
- unit: "associations"
- chart_type: line
- dimensions:
- - name: established
- - name: sctp.transitions
- description: SCTP Association Transitions
- unit: "transitions/s"
- chart_type: line
- dimensions:
- - name: active
- - name: passive
- - name: aborted
- - name: shutdown
- - name: sctp.packets
- description: SCTP Packets
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: sent
- - name: sctp.packet_errors
- description: SCTP Packet Errors
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: invalid
- - name: checksum
- - name: sctp.fragmentation
- description: SCTP Fragmentation
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: reassembled
- - name: fragmented
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/stat/nf_conntrack
- monitored_instance:
- name: Conntrack
- link: ""
- categories:
- - data-collection.linux-systems.firewall-metrics
- icon_filename: "firewall.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - connection tracking mechanism
- - netfilter
- - conntrack
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration monitors the connection tracking mechanism of Netfilter in the Linux Kernel."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: netfilter_conntrack_full
- link: https://github.com/netdata/netdata/blob/master/health/health.d/netfilter.conf
- metric: netfilter.conntrack_sockets
- info: netfilter connection tracker table size utilization
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: netfilter.conntrack_sockets
- description: Connection Tracker Connections
- unit: "active connections"
- chart_type: line
- dimensions:
- - name: connections
- - name: netfilter.conntrack_new
- description: Connection Tracker New Connections
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: new
- - name: ignore
- - name: invalid
- - name: netfilter.conntrack_changes
- description: Connection Tracker Changes
- unit: "changes/s"
- chart_type: line
- dimensions:
- - name: inserted
- - name: deleted
- - name: delete_list
- - name: netfilter.conntrack_expect
- description: Connection Tracker Expectations
- unit: "expectations/s"
- chart_type: line
- dimensions:
- - name: created
- - name: deleted
- - name: new
- - name: netfilter.conntrack_search
- description: Connection Tracker Searches
- unit: "searches/s"
- chart_type: line
- dimensions:
- - name: searched
- - name: restarted
- - name: found
- - name: netfilter.conntrack_errors
- description: Connection Tracker Errors
- unit: "events/s"
- chart_type: line
- dimensions:
- - name: icmp_error
- - name: error_failed
- - name: drop
- - name: early_drop
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/net/stat/synproxy
- monitored_instance:
- name: Synproxy
- link: ""
- categories:
- - data-collection.linux-systems.firewall-metrics
- icon_filename: "firewall.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - synproxy
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration provides statistics about the Synproxy netfilter module."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: netfilter.synproxy_syn_received
- description: SYNPROXY SYN Packets received
- unit: "packets/s"
- chart_type: line
- dimensions:
- - name: received
- - name: netfilter.synproxy_conn_reopened
- description: SYNPROXY Connections Reopened
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: reopened
- - name: netfilter.synproxy_cookies
- description: SYNPROXY TCP Cookies
- unit: "cookies/s"
- chart_type: line
- dimensions:
- - name: valid
- - name: invalid
- - name: retransmits
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/spl/kstat/zfs
- monitored_instance:
- name: ZFS Pools
- link: ""
- categories:
- - data-collection.linux-systems.filesystem-metrics.zfs
- icon_filename: "filesystem.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - zfs pools
- - pools
- - zfs
- - filesystem
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration provides metrics about the state of ZFS pools."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: zfs_pool_state_warn
- link: https://github.com/netdata/netdata/blob/master/health/health.d/zfs.conf
- metric: zfspool.state
- info: ZFS pool ${label:pool} state is degraded
- - name: zfs_pool_state_crit
- link: https://github.com/netdata/netdata/blob/master/health/health.d/zfs.conf
- metric: zfspool.state
- info: ZFS pool ${label:pool} state is faulted or unavail
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: zfs pool
- description: ""
- labels:
- - name: pool
- description: TBD
- metrics:
- - name: zfspool.state
- description: ZFS pool state
- unit: "boolean"
- chart_type: line
- dimensions:
- - name: online
- - name: degraded
- - name: faulted
- - name: offline
- - name: removed
- - name: unavail
- - name: suspended
- - meta:
- plugin_name: proc.plugin
- module_name: /proc/spl/kstat/zfs/arcstats
- monitored_instance:
- name: ZFS Adaptive Replacement Cache
- link: ""
- categories:
- - data-collection.linux-systems.filesystem-metrics.zfs
- icon_filename: "filesystem.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - zfs arc
- - arc
- - zfs
- - filesystem
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration monitors ZFS Adadptive Replacement Cache (ARC) statistics."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: zfs_memory_throttle
- link: https://github.com/netdata/netdata/blob/master/health/health.d/zfs.conf
- metric: zfs.memory_ops
- info: number of times ZFS had to limit the ARC growth in the last 10 minutes
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: ""
- labels: []
- metrics:
- - name: zfs.arc_size
- description: ZFS ARC Size
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: arcsz
- - name: target
- - name: min
- - name: max
- - name: zfs.l2_size
- description: ZFS L2 ARC Size
- unit: "MiB"
- chart_type: area
- dimensions:
- - name: actual
- - name: size
- - name: zfs.reads
- description: ZFS Reads
- unit: "reads/s"
- chart_type: area
- dimensions:
- - name: arc
- - name: demand
- - name: prefetch
- - name: metadata
- - name: l2
- - name: zfs.bytes
- description: ZFS ARC L2 Read/Write Rate
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: zfs.hits
- description: ZFS ARC Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.hits_rate
- description: ZFS ARC Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.dhits
- description: ZFS Demand Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.dhits_rate
- description: ZFS Demand Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.phits
- description: ZFS Prefetch Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.phits_rate
- description: ZFS Prefetch Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.mhits
- description: ZFS Metadata Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.mhits_rate
- description: ZFS Metadata Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.l2hits
- description: ZFS L2 Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.l2hits_rate
- description: ZFS L2 Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.list_hits
- description: ZFS List Hits
- unit: "hits/s"
- chart_type: area
- dimensions:
- - name: mfu
- - name: mfu_ghost
- - name: mru
- - name: mru_ghost
- - name: zfs.arc_size_breakdown
- description: ZFS ARC Size Breakdown
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: recent
- - name: frequent
- - name: zfs.memory_ops
- description: ZFS Memory Operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: direct
- - name: throttled
- - name: indirect
- - name: zfs.important_ops
- description: ZFS Important Operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: evict_skip
- - name: deleted
- - name: mutex_miss
- - name: hash_collisions
- - name: zfs.actual_hits
- description: ZFS Actual Cache Hits
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.actual_hits_rate
- description: ZFS Actual Cache Hits Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.demand_data_hits
- description: ZFS Data Demand Efficiency
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.demand_data_hits_rate
- description: ZFS Data Demand Efficiency Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.prefetch_data_hits
- description: ZFS Data Prefetch Efficiency
- unit: "percentage"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.prefetch_data_hits_rate
- description: ZFS Data Prefetch Efficiency Rate
- unit: "events/s"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: zfs.hash_elements
- description: ZFS ARC Hash Elements
- unit: "elements"
- chart_type: line
- dimensions:
- - name: current
- - name: max
- - name: zfs.hash_chains
- description: ZFS ARC Hash Chains
- unit: "chains"
- chart_type: line
- dimensions:
- - name: current
- - name: max
- - meta:
- plugin_name: proc.plugin
- module_name: /sys/fs/btrfs
- monitored_instance:
- name: BTRFS
- link: ""
- categories:
- - data-collection.linux-systems.filesystem-metrics.btrfs
- icon_filename: "filesystem.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - btrfs
- - filesystem
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration provides usage and error statistics from the BTRFS filesystem."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: btrfs_allocated
- link: https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf
- metric: btrfs.disk
- info: percentage of allocated BTRFS physical disk space
- os: "*"
- - name: btrfs_data
- link: https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf
- metric: btrfs.data
- info: utilization of BTRFS data space
- os: "*"
- - name: btrfs_metadata
- link: https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf
- metric: btrfs.metadata
- info: utilization of BTRFS metadata space
- os: "*"
- - name: btrfs_system
- link: https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf
- metric: btrfs.system
- info: utilization of BTRFS system space
- os: "*"
- - name: btrfs_device_read_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf
- metric: btrfs.device_errors
- info: number of encountered BTRFS read errors
- os: "*"
- - name: btrfs_device_write_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf
- metric: btrfs.device_errors
- info: number of encountered BTRFS write errors
- os: "*"
- - name: btrfs_device_flush_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf
- metric: btrfs.device_errors
- info: number of encountered BTRFS flush errors
- os: "*"
- - name: btrfs_device_corruption_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf
- metric: btrfs.device_errors
- info: number of encountered BTRFS corruption errors
- os: "*"
- - name: btrfs_device_generation_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/btrfs.conf
- metric: btrfs.device_errors
- info: number of encountered BTRFS generation errors
- os: "*"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: btrfs filesystem
- description: ""
- labels:
- - name: filesystem_uuid
- description: TBD
- - name: filesystem_label
- description: TBD
- metrics:
- - name: btrfs.disk
- description: BTRFS Physical Disk Allocation
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: unallocated
- - name: data_free
- - name: data_used
- - name: meta_free
- - name: meta_used
- - name: sys_free
- - name: sys_used
- - name: btrfs.data
- description: BTRFS Data Allocation
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: free
- - name: used
- - name: btrfs.metadata
- description: BTRFS Metadata Allocation
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: free
- - name: used
- - name: reserved
- - name: btrfs.system
- description: BTRFS System Allocation
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: free
- - name: used
- - name: btrfs.commits
- description: BTRFS Commits
- unit: "commits"
- chart_type: line
- dimensions:
- - name: commits
- - name: btrfs.commits_perc_time
- description: BTRFS Commits Time Share
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: commits
- - name: btrfs.commit_timings
- description: BTRFS Commit Timings
- unit: "ms"
- chart_type: line
- dimensions:
- - name: last
- - name: max
- - name: btrfs device
- description: ""
- labels:
- - name: device_id
- description: TBD
- - name: filesystem_uuid
- description: TBD
- - name: filesystem_label
- description: TBD
- metrics:
- - name: btrfs.device_errors
- description: BTRFS Device Errors
- unit: "errors"
- chart_type: line
- dimensions:
- - name: write_errs
- - name: read_errs
- - name: flush_errs
- - name: corruption_errs
- - name: generation_errs
- - meta:
- plugin_name: proc.plugin
- module_name: /sys/class/power_supply
- monitored_instance:
- name: Power Supply
- link: ""
- categories:
- - data-collection.linux-systems.power-supply-metrics
- icon_filename: "powersupply.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - psu
- - power supply
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration monitors Power supply metrics, such as battery status, AC power status and more."
- method_description: ""
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: linux_power_supply_capacity
- link: https://github.com/netdata/netdata/blob/master/health/health.d/linux_power_supply.conf
- metric: powersupply.capacity
- info: percentage of remaining power supply capacity
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: power device
- description: ""
- labels:
- - name: device
- description: TBD
- metrics:
- - name: powersupply.capacity
- description: Battery capacity
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: capacity
- - name: powersupply.charge
- description: Battery charge
- unit: "Ah"
- chart_type: line
- dimensions:
- - name: empty_design
- - name: empty
- - name: now
- - name: full
- - name: full_design
- - name: powersupply.energy
- description: Battery energy
- unit: "Wh"
- chart_type: line
- dimensions:
- - name: empty_design
- - name: empty
- - name: now
- - name: full
- - name: full_design
- - name: powersupply.voltage
- description: Power supply voltage
- unit: "V"
- chart_type: line
- dimensions:
- - name: min_design
- - name: min
- - name: now
- - name: max
- - name: max_design
- - meta:
- plugin_name: proc.plugin
- module_name: /sys/class/drm
- monitored_instance:
- name: AMD GPU
- link: "https://www.amd.com"
- categories:
- - data-collection.hardware-devices-and-sensors
- icon_filename: amd.svg
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - amd
- - gpu
- - hardware
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This integration monitors AMD GPU metrics, such as utilization, clock frequency and memory usage."
- method_description: "It reads `/sys/class/drm` to collect metrics for every AMD GPU card instance it encounters."
- supported_platforms:
- include:
- - Linux
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: ""
- description: ""
- options:
- description: ""
- folding:
- title: ""
- enabled: true
- list: []
- examples:
- folding:
- enabled: true
- title: ""
- list: []
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: gpu
- description: "These metrics refer to the GPU."
- labels:
- - name: product_name
- description: GPU product name (e.g. AMD RX 6600)
- metrics:
- - name: amdgpu.gpu_utilization
- description: GPU utilization
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: utilization
- - name: amdgpu.gpu_mem_utilization
- description: GPU memory utilization
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: utilization
- - name: amdgpu.gpu_clk_frequency
- description: GPU clock frequency
- unit: "MHz"
- chart_type: line
- dimensions:
- - name: frequency
- - name: amdgpu.gpu_mem_clk_frequency
- description: GPU memory clock frequency
- unit: "MHz"
- chart_type: line
- dimensions:
- - name: frequency
- - name: amdgpu.gpu_mem_vram_usage_perc
- description: VRAM memory usage percentage
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: usage
- - name: amdgpu.gpu_mem_vram_usage
- description: VRAM memory usage
- unit: "bytes"
- chart_type: area
- dimensions:
- - name: free
- - name: used
- - name: amdgpu.gpu_mem_vis_vram_usage_perc
- description: visible VRAM memory usage percentage
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: usage
- - name: amdgpu.gpu_mem_vis_vram_usage
- description: visible VRAM memory usage
- unit: "bytes"
- chart_type: area
- dimensions:
- - name: free
- - name: used
- - name: amdgpu.gpu_mem_gtt_usage_perc
- description: GTT memory usage percentage
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: usage
- - name: amdgpu.gpu_mem_gtt_usage
- description: GTT memory usage
- unit: "bytes"
- chart_type: area
- dimensions:
- - name: free
- - name: used
diff --git a/collectors/profile.plugin/Makefile.am b/collectors/profile.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/profile.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/python.d.plugin/Makefile.am b/collectors/python.d.plugin/Makefile.am
deleted file mode 100644
index ca49c1c02..000000000
--- a/collectors/python.d.plugin/Makefile.am
+++ /dev/null
@@ -1,229 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-CLEANFILES = \
- python.d.plugin \
- $(NULL)
-
-include $(top_srcdir)/build/subst.inc
-SUFFIXES = .in
-
-dist_libconfig_DATA = \
- python.d.conf \
- $(NULL)
-
-dist_plugins_SCRIPTS = \
- python.d.plugin \
- $(NULL)
-
-dist_noinst_DATA = \
- python.d.plugin.in \
- README.md \
- $(NULL)
-
-dist_python_SCRIPTS = \
- $(NULL)
-
-dist_python_DATA = \
- $(NULL)
-
-userpythonconfigdir=$(configdir)/python.d
-dist_userpythonconfig_DATA = \
- $(NULL)
-
-# Explicitly install directories to avoid permission issues due to umask
-install-exec-local:
- $(INSTALL) -d $(DESTDIR)$(userpythonconfigdir)
-
-pythonconfigdir=$(libconfigdir)/python.d
-dist_pythonconfig_DATA = \
- $(NULL)
-
-include adaptec_raid/Makefile.inc
-include alarms/Makefile.inc
-include am2320/Makefile.inc
-include anomalies/Makefile.inc
-include beanstalk/Makefile.inc
-include bind_rndc/Makefile.inc
-include boinc/Makefile.inc
-include ceph/Makefile.inc
-include changefinder/Makefile.inc
-include dovecot/Makefile.inc
-include example/Makefile.inc
-include exim/Makefile.inc
-include fail2ban/Makefile.inc
-include gearman/Makefile.inc
-include go_expvar/Makefile.inc
-include haproxy/Makefile.inc
-include hddtemp/Makefile.inc
-include hpssa/Makefile.inc
-include icecast/Makefile.inc
-include ipfs/Makefile.inc
-include litespeed/Makefile.inc
-include megacli/Makefile.inc
-include memcached/Makefile.inc
-include monit/Makefile.inc
-include nvidia_smi/Makefile.inc
-include nsd/Makefile.inc
-include openldap/Makefile.inc
-include oracledb/Makefile.inc
-include pandas/Makefile.inc
-include postfix/Makefile.inc
-include puppet/Makefile.inc
-include rethinkdbs/Makefile.inc
-include retroshare/Makefile.inc
-include riakkv/Makefile.inc
-include samba/Makefile.inc
-include sensors/Makefile.inc
-include smartd_log/Makefile.inc
-include spigotmc/Makefile.inc
-include squid/Makefile.inc
-include tomcat/Makefile.inc
-include tor/Makefile.inc
-include traefik/Makefile.inc
-include uwsgi/Makefile.inc
-include varnish/Makefile.inc
-include w1sensor/Makefile.inc
-include zscores/Makefile.inc
-
-pythonmodulesdir=$(pythondir)/python_modules
-dist_pythonmodules_DATA = \
- python_modules/__init__.py \
- $(NULL)
-
-basesdir=$(pythonmodulesdir)/bases
-dist_bases_DATA = \
- python_modules/bases/__init__.py \
- python_modules/bases/charts.py \
- python_modules/bases/collection.py \
- python_modules/bases/loaders.py \
- python_modules/bases/loggers.py \
- $(NULL)
-
-bases_framework_servicesdir=$(basesdir)/FrameworkServices
-dist_bases_framework_services_DATA = \
- python_modules/bases/FrameworkServices/__init__.py \
- python_modules/bases/FrameworkServices/ExecutableService.py \
- python_modules/bases/FrameworkServices/LogService.py \
- python_modules/bases/FrameworkServices/MySQLService.py \
- python_modules/bases/FrameworkServices/SimpleService.py \
- python_modules/bases/FrameworkServices/SocketService.py \
- python_modules/bases/FrameworkServices/UrlService.py \
- $(NULL)
-
-third_partydir=$(pythonmodulesdir)/third_party
-dist_third_party_DATA = \
- python_modules/third_party/__init__.py \
- python_modules/third_party/ordereddict.py \
- python_modules/third_party/lm_sensors.py \
- python_modules/third_party/mcrcon.py \
- python_modules/third_party/boinc_client.py \
- python_modules/third_party/monotonic.py \
- python_modules/third_party/filelock.py \
- $(NULL)
-
-pythonyaml2dir=$(pythonmodulesdir)/pyyaml2
-dist_pythonyaml2_DATA = \
- python_modules/pyyaml2/__init__.py \
- python_modules/pyyaml2/composer.py \
- python_modules/pyyaml2/constructor.py \
- python_modules/pyyaml2/cyaml.py \
- python_modules/pyyaml2/dumper.py \
- python_modules/pyyaml2/emitter.py \
- python_modules/pyyaml2/error.py \
- python_modules/pyyaml2/events.py \
- python_modules/pyyaml2/loader.py \
- python_modules/pyyaml2/nodes.py \
- python_modules/pyyaml2/parser.py \
- python_modules/pyyaml2/reader.py \
- python_modules/pyyaml2/representer.py \
- python_modules/pyyaml2/resolver.py \
- python_modules/pyyaml2/scanner.py \
- python_modules/pyyaml2/serializer.py \
- python_modules/pyyaml2/tokens.py \
- $(NULL)
-
-pythonyaml3dir=$(pythonmodulesdir)/pyyaml3
-dist_pythonyaml3_DATA = \
- python_modules/pyyaml3/__init__.py \
- python_modules/pyyaml3/composer.py \
- python_modules/pyyaml3/constructor.py \
- python_modules/pyyaml3/cyaml.py \
- python_modules/pyyaml3/dumper.py \
- python_modules/pyyaml3/emitter.py \
- python_modules/pyyaml3/error.py \
- python_modules/pyyaml3/events.py \
- python_modules/pyyaml3/loader.py \
- python_modules/pyyaml3/nodes.py \
- python_modules/pyyaml3/parser.py \
- python_modules/pyyaml3/reader.py \
- python_modules/pyyaml3/representer.py \
- python_modules/pyyaml3/resolver.py \
- python_modules/pyyaml3/scanner.py \
- python_modules/pyyaml3/serializer.py \
- python_modules/pyyaml3/tokens.py \
- $(NULL)
-
-python_urllib3dir=$(pythonmodulesdir)/urllib3
-dist_python_urllib3_DATA = \
- python_modules/urllib3/__init__.py \
- python_modules/urllib3/_collections.py \
- python_modules/urllib3/connection.py \
- python_modules/urllib3/connectionpool.py \
- python_modules/urllib3/exceptions.py \
- python_modules/urllib3/fields.py \
- python_modules/urllib3/filepost.py \
- python_modules/urllib3/response.py \
- python_modules/urllib3/poolmanager.py \
- python_modules/urllib3/request.py \
- $(NULL)
-
-python_urllib3_utildir=$(python_urllib3dir)/util
-dist_python_urllib3_util_DATA = \
- python_modules/urllib3/util/__init__.py \
- python_modules/urllib3/util/connection.py \
- python_modules/urllib3/util/request.py \
- python_modules/urllib3/util/response.py \
- python_modules/urllib3/util/retry.py \
- python_modules/urllib3/util/selectors.py \
- python_modules/urllib3/util/ssl_.py \
- python_modules/urllib3/util/timeout.py \
- python_modules/urllib3/util/url.py \
- python_modules/urllib3/util/wait.py \
- $(NULL)
-
-python_urllib3_packagesdir=$(python_urllib3dir)/packages
-dist_python_urllib3_packages_DATA = \
- python_modules/urllib3/packages/__init__.py \
- python_modules/urllib3/packages/ordered_dict.py \
- python_modules/urllib3/packages/six.py \
- $(NULL)
-
-python_urllib3_backportsdir=$(python_urllib3_packagesdir)/backports
-dist_python_urllib3_backports_DATA = \
- python_modules/urllib3/packages/backports/__init__.py \
- python_modules/urllib3/packages/backports/makefile.py \
- $(NULL)
-
-python_urllib3_ssl_match_hostnamedir=$(python_urllib3_packagesdir)/ssl_match_hostname
-dist_python_urllib3_ssl_match_hostname_DATA = \
- python_modules/urllib3/packages/ssl_match_hostname/__init__.py \
- python_modules/urllib3/packages/ssl_match_hostname/_implementation.py \
- $(NULL)
-
-python_urllib3_contribdir=$(python_urllib3dir)/contrib
-dist_python_urllib3_contrib_DATA = \
- python_modules/urllib3/contrib/__init__.py \
- python_modules/urllib3/contrib/appengine.py \
- python_modules/urllib3/contrib/ntlmpool.py \
- python_modules/urllib3/contrib/pyopenssl.py \
- python_modules/urllib3/contrib/securetransport.py \
- python_modules/urllib3/contrib/socks.py \
- $(NULL)
-
-python_urllib3_securetransportdir=$(python_urllib3_contribdir)/_securetransport
-dist_python_urllib3_securetransport_DATA = \
- python_modules/urllib3/contrib/_securetransport/__init__.py \
- python_modules/urllib3/contrib/_securetransport/bindings.py \
- python_modules/urllib3/contrib/_securetransport/low_level.py \
- $(NULL)
diff --git a/collectors/python.d.plugin/README.md b/collectors/python.d.plugin/README.md
deleted file mode 100644
index 569543d16..000000000
--- a/collectors/python.d.plugin/README.md
+++ /dev/null
@@ -1,77 +0,0 @@
-<!--
-title: "python.d.plugin"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/README.md"
-sidebar_label: "python.d.plugin"
-learn_status: "Published"
-learn_topic_type: "Tasks"
-learn_rel_path: "Developers/External plugins/python.d.plugin"
--->
-
-# python.d.plugin
-
-`python.d.plugin` is a Netdata external plugin. It is an **orchestrator** for data collection modules written in `python`.
-
-1. It runs as an independent process `ps fax` shows it
-2. It is started and stopped automatically by Netdata
-3. It communicates with Netdata via a unidirectional pipe (sending data to the `netdata` daemon)
-4. Supports any number of data collection **modules**
-5. Allows each **module** to have one or more data collection **jobs**
-6. Each **job** is collecting one or more metrics from a single data source
-
-## Disclaimer
-
-All third party libraries should be installed system-wide or in `python_modules` directory.
-Module configurations are written in YAML and **pyYAML is required**.
-
-Every configuration file must have one of two formats:
-
-- Configuration for only one job:
-
-```yaml
-update_every : 2 # update frequency
-priority : 20000 # where it is shown on dashboard
-
-other_var1 : bla # variables passed to module
-other_var2 : alb
-```
-
-- Configuration for many jobs (ex. mysql):
-
-```yaml
-# module defaults:
-update_every : 2
-priority : 20000
-
-local: # job name
- update_every : 5 # job update frequency
- other_var1 : some_val # module specific variable
-
-other_job:
- priority : 5 # job position on dashboard
- other_var2 : val # module specific variable
-```
-
-`update_every` and `priority` are always optional.
-
-## How to debug a python module
-
-```
-# become user netdata
-sudo su -s /bin/bash netdata
-```
-
-Depending on where Netdata was installed, execute one of the following commands to trace the execution of a python module:
-
-```
-# execute the plugin in debug mode, for a specific module
-/opt/netdata/usr/libexec/netdata/plugins.d/python.d.plugin <module> debug trace
-/usr/libexec/netdata/plugins.d/python.d.plugin <module> debug trace
-```
-
-Where `[module]` is the directory name under <https://github.com/netdata/netdata/tree/master/collectors/python.d.plugin>
-
-**Note**: If you would like execute a collector in debug mode while it is still running by Netdata, you can pass the `nolock` CLI option to the above commands.
-
-## How to write a new module
-
-See [develop a custom collector in Python](https://github.com/netdata/netdata/edit/master/docs/guides/python-collector.md).
diff --git a/collectors/python.d.plugin/adaptec_raid/Makefile.inc b/collectors/python.d.plugin/adaptec_raid/Makefile.inc
deleted file mode 100644
index 716cdb235..000000000
--- a/collectors/python.d.plugin/adaptec_raid/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += adaptec_raid/adaptec_raid.chart.py
-dist_pythonconfig_DATA += adaptec_raid/adaptec_raid.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += adaptec_raid/README.md adaptec_raid/Makefile.inc
-
diff --git a/collectors/python.d.plugin/adaptec_raid/metadata.yaml b/collectors/python.d.plugin/adaptec_raid/metadata.yaml
deleted file mode 100644
index c69baff4a..000000000
--- a/collectors/python.d.plugin/adaptec_raid/metadata.yaml
+++ /dev/null
@@ -1,167 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: adaptec_raid
- monitored_instance:
- name: AdaptecRAID
- link: "https://www.microchip.com/en-us/products/storage"
- categories:
- - data-collection.storage-mount-points-and-filesystems
- icon_filename: "adaptec.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - storage
- - raid-controller
- - manage-disks
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- This collector monitors Adaptec RAID hardware storage controller metrics about both physical and logical drives.
- method_description: |
- It uses the arcconf command line utility (from adaptec) to monitor your raid controller.
-
- Executed commands:
- - `sudo -n arcconf GETCONFIG 1 LD`
- - `sudo -n arcconf GETCONFIG 1 PD`
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: "The module uses arcconf, which can only be executed by root. It uses sudo and assumes that it is configured such that the netdata user can execute arcconf as root without a password."
- default_behavior:
- auto_detection:
- description: "After all the permissions are satisfied, netdata should be to execute commands via the arcconf command line utility"
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: Grant permissions for netdata, to run arcconf as sudoer
- description: |
- The module uses arcconf, which can only be executed by root. It uses sudo and assumes that it is configured such that the netdata user can execute arcconf as root without a password.
-
- Add to your /etc/sudoers file:
- which arcconf shows the full path to the binary.
-
- ```bash
- netdata ALL=(root) NOPASSWD: /path/to/arcconf
- ```
- - title: Reset Netdata's systemd unit CapabilityBoundingSet (Linux distributions with systemd)
- description: |
- The default CapabilityBoundingSet doesn't allow using sudo, and is quite strict in general. Resetting is not optimal, but a next-best solution given the inability to execute arcconf using sudo.
-
- As root user, do the following:
-
- ```bash
- mkdir /etc/systemd/system/netdata.service.d
- echo -e '[Service]\nCapabilityBoundingSet=~' | tee /etc/systemd/system/netdata.service.d/unset-capability-bounding-set.conf
- systemctl daemon-reload
- systemctl restart netdata.service
- ```
- configuration:
- file:
- name: "python.d/adaptec_raid.conf"
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Basic
- folding:
- enabled: false
- description: A basic example configuration per job
- config: |
- job_name:
- name: my_job_name
- update_every: 1 # the JOB's data collection frequency
- priority: 60000 # the JOB's order on the dashboard
- penalty: yes # the JOB's penalty
- autodetection_retry: 0 # the JOB's re-check interval in seconds
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: adaptec_raid_ld_status
- link: https://github.com/netdata/netdata/blob/master/health/health.d/adaptec_raid.conf
- metric: adaptec_raid.ld_status
- info: logical device status is failed or degraded
- - name: adaptec_raid_pd_state
- link: https://github.com/netdata/netdata/blob/master/health/health.d/adaptec_raid.conf
- metric: adaptec_raid.pd_state
- info: physical device state is not online
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: adaptec_raid.ld_status
- description: "Status of logical devices (1: Failed or Degraded)"
- unit: "bool"
- chart_type: line
- dimensions:
- - name: a dimension per logical device
- - name: adaptec_raid.pd_state
- description: "State of physical devices (1: not Online)"
- unit: "bool"
- chart_type: line
- dimensions:
- - name: a dimension per physical device
- - name: adaptec_raid.smart_warnings
- description: S.M.A.R.T warnings
- unit: "count"
- chart_type: line
- dimensions:
- - name: a dimension per physical device
- - name: adaptec_raid.temperature
- description: Temperature
- unit: "celsius"
- chart_type: line
- dimensions:
- - name: a dimension per physical device
diff --git a/collectors/python.d.plugin/alarms/Makefile.inc b/collectors/python.d.plugin/alarms/Makefile.inc
deleted file mode 100644
index c2de11724..000000000
--- a/collectors/python.d.plugin/alarms/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += alarms/alarms.chart.py
-dist_pythonconfig_DATA += alarms/alarms.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += alarms/README.md alarms/Makefile.inc
-
diff --git a/collectors/python.d.plugin/alarms/metadata.yaml b/collectors/python.d.plugin/alarms/metadata.yaml
deleted file mode 100644
index 30a897787..000000000
--- a/collectors/python.d.plugin/alarms/metadata.yaml
+++ /dev/null
@@ -1,177 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: alarms
- monitored_instance:
- name: Netdata Agent alarms
- link: https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/alarms/README.md
- categories:
- - data-collection.other
- icon_filename: ""
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - alarms
- - netdata
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- This collector creates an 'Alarms' menu with one line plot of `alarms.status`.
- method_description: |
- Alarm status is read from the Netdata agent rest api [`/api/v1/alarms?all`](https://learn.netdata.cloud/api#/alerts/alerts1).
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: |
- It discovers instances of Netdata running on localhost, and gathers metrics from `http://127.0.0.1:19999/api/v1/alarms?all`. `CLEAR` status is mapped to `0`, `WARNING` to `1` and `CRITICAL` to `2`. Also, by default all alarms produced will be monitored.
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: python.d/alarms.conf
- description: ""
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: Config options
- enabled: true
- list:
- - name: url
- description: Netdata agent alarms endpoint to collect from. Can be local or remote so long as reachable by agent.
- default_value: http://127.0.0.1:19999/api/v1/alarms?all
- required: true
- - name: status_map
- description: Mapping of alarm status to integer number that will be the metric value collected.
- default_value: '{"CLEAR": 0, "WARNING": 1, "CRITICAL": 2}'
- required: true
- - name: collect_alarm_values
- description: set to true to include a chart with calculated alarm values over time.
- default_value: false
- required: true
- - name: alarm_status_chart_type
- description: define the type of chart for plotting status over time e.g. 'line' or 'stacked'.
- default_value: "line"
- required: true
- - name: alarm_contains_words
- description: >
- A "," separated list of words you want to filter alarm names for. For example 'cpu,load' would filter for only alarms with "cpu" or "load" in alarm name. Default includes all.
- default_value: ""
- required: true
- - name: alarm_excludes_words
- description: >
- A "," separated list of words you want to exclude based on alarm name. For example 'cpu,load' would exclude all alarms with "cpu" or "load" in alarm name. Default excludes None.
- default_value: ""
- required: true
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 10
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ""
- required: false
- examples:
- folding:
- enabled: true
- title: Config
- list:
- - name: Basic
- folding:
- enabled: false
- description: A basic example configuration.
- config: |
- jobs:
- url: 'http://127.0.0.1:19999/api/v1/alarms?all'
- - name: Advanced
- folding:
- enabled: true
- description: |
- An advanced example configuration with multiple jobs collecting different subsets of alarms for plotting on different charts.
- "ML" job will collect status and values for all alarms with "ml_" in the name. Default job will collect status for all other alarms.
- config: |
- ML:
- update_every: 5
- url: 'http://127.0.0.1:19999/api/v1/alarms?all'
- status_map:
- CLEAR: 0
- WARNING: 1
- CRITICAL: 2
- collect_alarm_values: true
- alarm_status_chart_type: 'stacked'
- alarm_contains_words: 'ml_'
-
- Default:
- update_every: 5
- url: 'http://127.0.0.1:19999/api/v1/alarms?all'
- status_map:
- CLEAR: 0
- WARNING: 1
- CRITICAL: 2
- collect_alarm_values: false
- alarm_status_chart_type: 'stacked'
- alarm_excludes_words: 'ml_'
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: |
- These metrics refer to the entire monitored application.
- labels: []
- metrics:
- - name: alarms.status
- description: Alarms ({status mapping})
- unit: "status"
- chart_type: line
- dimensions:
- - name: a dimension per alarm representing the latest status of the alarm.
- - name: alarms.values
- description: Alarm Values
- unit: "value"
- chart_type: line
- dimensions:
- - name: a dimension per alarm representing the latest collected value of the alarm.
diff --git a/collectors/python.d.plugin/am2320/Makefile.inc b/collectors/python.d.plugin/am2320/Makefile.inc
deleted file mode 100644
index 48e5a8892..000000000
--- a/collectors/python.d.plugin/am2320/Makefile.inc
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# install these files
-dist_python_DATA += am2320/am2320.chart.py
-dist_pythonconfig_DATA += am2320/am2320.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += am2320/README.md am2320/Makefile.inc
diff --git a/collectors/python.d.plugin/anomalies/Makefile.inc b/collectors/python.d.plugin/anomalies/Makefile.inc
deleted file mode 100644
index 94937b36a..000000000
--- a/collectors/python.d.plugin/anomalies/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += anomalies/anomalies.chart.py
-dist_pythonconfig_DATA += anomalies/anomalies.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += anomalies/README.md anomalies/Makefile.inc
-
diff --git a/collectors/python.d.plugin/anomalies/README.md b/collectors/python.d.plugin/anomalies/README.md
deleted file mode 100644
index 80f505375..000000000
--- a/collectors/python.d.plugin/anomalies/README.md
+++ /dev/null
@@ -1,248 +0,0 @@
-<!--
-title: "Anomaly detection with Netdata"
-description: "Use ML-driven anomaly detection to narrow your focus to only affected metrics and services/processes on your node to shorten root cause analysis."
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/anomalies/README.md"
-sidebar_url: "Anomalies"
-sidebar_label: "anomalies"
-learn_status: "Published"
-learn_rel_path: "Integrations/Monitor/Anything"
--->
-
-# Anomaly detection with Netdata
-
-**Note**: Check out the [Netdata Anomaly Advisor](https://github.com/netdata/netdata/blob/master/docs/cloud/insights/anomaly-advisor.md) for a more native anomaly detection experience within Netdata.
-
-This collector uses the Python [PyOD](https://pyod.readthedocs.io/en/latest/index.html) library to perform unsupervised [anomaly detection](https://en.wikipedia.org/wiki/Anomaly_detection) on your Netdata charts and/or dimensions.
-
-Instead of this collector just _collecting_ data, it also does some computation on the data it collects to return an anomaly probability and anomaly flag for each chart or custom model you define. This computation consists of a **train** function that runs every `train_n_secs` to train the ML models to learn what 'normal' typically looks like on your node. At each iteration there is also a **predict** function that uses the latest trained models and most recent metrics to produce an anomaly probability and anomaly flag for each chart or custom model you define.
-
-> As this is a somewhat unique collector and involves often subjective concepts like anomalies and anomaly probabilities, we would love to hear any feedback on it from the community. Please let us know on the [community forum](https://community.netdata.cloud/t/anomalies-collector-feedback-megathread/767) or drop us a note at [analytics-ml-team@netdata.cloud](mailto:analytics-ml-team@netdata.cloud) for any and all feedback, both positive and negative. This sort of feedback is priceless to help us make complex features more useful.
-
-## Charts
-
-Two charts are produced:
-
-- **Anomaly Probability** (`anomalies.probability`): This chart shows the probability that the latest observed data is anomalous based on the trained model for that chart (using the [`predict_proba()`](https://pyod.readthedocs.io/en/latest/api_cc.html#pyod.models.base.BaseDetector.predict_proba) method of the trained PyOD model).
-- **Anomaly** (`anomalies.anomaly`): This chart shows `1` or `0` predictions of if the latest observed data is considered anomalous or not based on the trained model (using the [`predict()`](https://pyod.readthedocs.io/en/latest/api_cc.html#pyod.models.base.BaseDetector.predict) method of the trained PyOD model).
-
-Below is an example of the charts produced by this collector and how they might look when things are 'normal' on the node. The anomaly probabilities tend to bounce randomly around a typically low probability range, one or two might randomly jump or drift outside of this range every now and then and show up as anomalies on the anomaly chart.
-
-![netdata-anomalies-collector-normal](https://user-images.githubusercontent.com/2178292/100663699-99755000-334e-11eb-922f-0c41a0176484.jpg)
-
-If we then go onto the system and run a command like `stress-ng --all 2` to create some [stress](https://wiki.ubuntu.com/Kernel/Reference/stress-ng), we see some charts begin to have anomaly probabilities that jump outside the typical range. When the anomaly probabilities change enough, we will start seeing anomalies being flagged on the `anomalies.anomaly` chart. The idea is that these charts are the most anomalous right now so could be a good place to start your troubleshooting.
-
-![netdata-anomalies-collector-abnormal](https://user-images.githubusercontent.com/2178292/100663710-9bd7aa00-334e-11eb-9d14-76fda73bc309.jpg)
-
-Then, as the issue passes, the anomaly probabilities should settle back down into their 'normal' range again.
-
-![netdata-anomalies-collector-normal-again](https://user-images.githubusercontent.com/2178292/100666681-481a9000-3351-11eb-9979-64728ee2dfb6.jpg)
-
-## Requirements
-
-- This collector will only work with Python 3 and requires the packages below be installed.
-- Typically you will not need to do this, but, if needed, to ensure Python 3 is used you can add the below line to the `[plugin:python.d]` section of `netdata.conf`
-
-```conf
-[plugin:python.d]
- # update every = 1
- command options = -ppython3
-```
-
-Install the required python libraries.
-
-```bash
-# become netdata user
-sudo su -s /bin/bash netdata
-# install required packages for the netdata user
-pip3 install --user netdata-pandas==0.0.38 numba==0.50.1 scikit-learn==0.23.2 pyod==0.8.3
-```
-
-## Configuration
-
-Install the Python requirements above, enable the collector and restart Netdata.
-
-```bash
-cd /etc/netdata/
-sudo ./edit-config python.d.conf
-# Set `anomalies: no` to `anomalies: yes`
-sudo systemctl restart netdata
-```
-
-The configuration for the anomalies collector defines how it will behave on your system and might take some experimentation with over time to set it optimally for your node. Out of the box, the config comes with some [sane defaults](https://www.netdata.cloud/blog/redefining-monitoring-netdata/) to get you started that try to balance the flexibility and power of the ML models with the goal of being as cheap as possible in term of cost on the node resources.
-
-_**Note**: If you are unsure about any of the below configuration options then it's best to just ignore all this and leave the `anomalies.conf` file alone to begin with. Then you can return to it later if you would like to tune things a bit more once the collector is running for a while and you have a feeling for its performance on your node._
-
-Edit the `python.d/anomalies.conf` configuration file using `edit-config` from the your agent's [config
-directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is usually at `/etc/netdata`.
-
-```bash
-cd /etc/netdata # Replace this path with your Netdata config directory, if different
-sudo ./edit-config python.d/anomalies.conf
-```
-
-The default configuration should look something like this. Here you can see each parameter (with sane defaults) and some information about each one and what it does.
-
-```conf
-# -
-# JOBS (data collection sources)
-
-# Pull data from local Netdata node.
-anomalies:
- name: 'Anomalies'
-
- # Host to pull data from.
- host: '127.0.0.1:19999'
-
- # Username and Password for Netdata if using basic auth.
- # username: '???'
- # password: '???'
-
- # Use http or https to pull data
- protocol: 'http'
-
- # SSL verify parameter for requests.get() calls
- tls_verify: true
-
- # What charts to pull data for - A regex like 'system\..*|' or 'system\..*|apps.cpu|apps.mem' etc.
- charts_regex: 'system\..*'
-
- # Charts to exclude, useful if you would like to exclude some specific charts.
- # Note: should be a ',' separated string like 'chart.name,chart.name'.
- charts_to_exclude: 'system.uptime,system.entropy'
-
- # What model to use - can be one of 'pca', 'hbos', 'iforest', 'cblof', 'loda', 'copod' or 'feature_bagging'.
- # More details here: https://pyod.readthedocs.io/en/latest/pyod.models.html.
- model: 'pca'
-
- # Max number of observations to train on, to help cap compute cost of training model if you set a very large train_n_secs.
- train_max_n: 100000
-
- # How often to re-train the model (assuming update_every=1 then train_every_n=1800 represents (re)training every 30 minutes).
- # Note: If you want to turn off re-training set train_every_n=0 and after initial training the models will not be retrained.
- train_every_n: 1800
-
- # The length of the window of data to train on (14400 = last 4 hours).
- train_n_secs: 14400
-
- # How many prediction steps after a train event to just use previous prediction value for.
- # Used to reduce possibility of the training step itself appearing as an anomaly on the charts.
- train_no_prediction_n: 10
-
- # If you would like to train the model for the first time on a specific window then you can define it using the below two variables.
- # Start of training data for initial model.
- # initial_train_data_after: 1604578857
-
- # End of training data for initial model.
- # initial_train_data_before: 1604593257
-
- # If you would like to ignore recent data in training then you can offset it by offset_n_secs.
- offset_n_secs: 0
-
- # How many lagged values of each dimension to include in the 'feature vector' each model is trained on.
- lags_n: 5
-
- # How much smoothing to apply to each dimension in the 'feature vector' each model is trained on.
- smooth_n: 3
-
- # How many differences to take in preprocessing your data.
- # More info on differencing here: https://en.wikipedia.org/wiki/Autoregressive_integrated_moving_average#Differencing
- # diffs_n=0 would mean training models on the raw values of each dimension.
- # diffs_n=1 means everything is done in terms of differences.
- diffs_n: 1
-
- # What is the typical proportion of anomalies in your data on average?
- # This parameter can control the sensitivity of your models to anomalies.
- # Some discussion here: https://github.com/yzhao062/pyod/issues/144
- contamination: 0.001
-
- # Set to true to include an "average_prob" dimension on anomalies probability chart which is
- # just the average of all anomaly probabilities at each time step
- include_average_prob: true
-
- # Define any custom models you would like to create anomaly probabilities for, some examples below to show how.
- # For example below example creates two custom models, one to run anomaly detection user and system cpu for our demo servers
- # and one on the cpu and mem apps metrics for the python.d.plugin.
- # custom_models:
- # - name: 'demos_cpu'
- # dimensions: 'london.my-netdata.io::system.cpu|user,london.my-netdata.io::system.cpu|system,newyork.my-netdata.io::system.cpu|user,newyork.my-netdata.io::system.cpu|system'
- # - name: 'apps_python_d_plugin'
- # dimensions: 'apps.cpu|python.d.plugin,apps.mem|python.d.plugin'
-
- # Set to true to normalize, using min-max standardization, features used for the custom models.
- # Useful if your custom models contain dimensions on very different scales an model you use does
- # not internally do its own normalization. Usually best to leave as false.
- # custom_models_normalize: false
-```
-
-## Custom models
-
-In the `anomalies.conf` file you can also define some "custom models" which you can use to group one or more metrics into a single model much like is done by default for the charts you specify. This is useful if you have a handful of metrics that exist in different charts but perhaps are related to the same underlying thing you would like to perform anomaly detection on, for example a specific app or user.
-
-To define a custom model you would include configuration like below in `anomalies.conf`. By default there should already be some commented out examples in there.
-
-`name` is a name you give your custom model, this is what will appear alongside any other specified charts in the `anomalies.probability` and `anomalies.anomaly` charts. `dimensions` is a string of metrics you want to include in your custom model. By default the [netdata-pandas](https://github.com/netdata/netdata-pandas) library used to pull the data from Netdata uses a "chart.a|dim.1" type of naming convention in the pandas columns it returns, hence the `dimensions` string should look like "chart.name|dimension.name,chart.name|dimension.name". The examples below hopefully make this clear.
-
-```yaml
-custom_models:
- # a model for anomaly detection on the netdata user in terms of cpu, mem, threads, processes and sockets.
- - name: 'user_netdata'
- dimensions: 'users.cpu|netdata,users.mem|netdata,users.threads|netdata,users.processes|netdata,users.sockets|netdata'
- # a model for anomaly detection on the netdata python.d.plugin app in terms of cpu, mem, threads, processes and sockets.
- - name: 'apps_python_d_plugin'
- dimensions: 'apps.cpu|python.d.plugin,apps.mem|python.d.plugin,apps.threads|python.d.plugin,apps.processes|python.d.plugin,apps.sockets|python.d.plugin'
-
-custom_models_normalize: false
-```
-
-## Troubleshooting
-
-To see any relevant log messages you can use a command like below.
-
-```bash
-`grep 'anomalies' /var/log/netdata/error.log`
-```
-
-If you would like to log in as `netdata` user and run the collector in debug mode to see more detail.
-
-```bash
-# become netdata user
-sudo su -s /bin/bash netdata
-# run collector in debug using `nolock` option if netdata is already running the collector itself.
-/usr/libexec/netdata/plugins.d/python.d.plugin anomalies debug trace nolock
-```
-
-## Deepdive tutorial
-
-If you would like to go deeper on what exactly the anomalies collector is doing under the hood then check out this [deepdive tutorial](https://github.com/netdata/community/blob/main/netdata-agent-api/netdata-pandas/anomalies_collector_deepdive.ipynb) in our community repo where you can play around with some data from our demo servers (or your own if its accessible to you) and work through the calculations step by step.
-
-(Note: as its a Jupyter Notebook it might render a little prettier on [nbviewer](https://nbviewer.jupyter.org/github/netdata/community/blob/main/netdata-agent-api/netdata-pandas/anomalies_collector_deepdive.ipynb))
-
-## Notes
-
-- Python 3 is required as the [`netdata-pandas`](https://github.com/netdata/netdata-pandas) package uses Python async libraries ([asks](https://pypi.org/project/asks/) and [trio](https://pypi.org/project/trio/)) to make asynchronous calls to the [Netdata REST API](https://github.com/netdata/netdata/blob/master/web/api/README.md) to get the required data for each chart.
-- Python 3 is also required for the underlying ML libraries of [numba](https://pypi.org/project/numba/), [scikit-learn](https://pypi.org/project/scikit-learn/), and [PyOD](https://pypi.org/project/pyod/).
-- It may take a few hours or so (depending on your choice of `train_secs_n`) for the collector to 'settle' into it's typical behaviour in terms of the trained models and probabilities you will see in the normal running of your node.
-- As this collector does most of the work in Python itself, with [PyOD](https://pyod.readthedocs.io/en/latest/) leveraging [numba](https://numba.pydata.org/) under the hood, you may want to try it out first on a test or development system to get a sense of its performance characteristics on a node similar to where you would like to use it.
-- `lags_n`, `smooth_n`, and `diffs_n` together define the preprocessing done to the raw data before models are trained and before each prediction. This essentially creates a [feature vector](https://en.wikipedia.org/wiki/Feature_(machine_learning)#:~:text=In%20pattern%20recognition%20and%20machine,features%20that%20represent%20some%20object.&text=Feature%20vectors%20are%20often%20combined,score%20for%20making%20a%20prediction.) for each chart model (or each custom model). The default settings for these parameters aim to create a rolling matrix of recent smoothed [differenced](https://en.wikipedia.org/wiki/Autoregressive_integrated_moving_average#Differencing) values for each chart. The aim of the model then is to score how unusual this 'matrix' of features is for each chart based on what it has learned as 'normal' from the training data. So as opposed to just looking at the single most recent value of a dimension and considering how strange it is, this approach looks at a recent smoothed window of all dimensions for a chart (or dimensions in a custom model) and asks how unusual the data as a whole looks. This should be more flexible in capturing a wider range of [anomaly types](https://andrewm4894.com/2020/10/19/different-types-of-time-series-anomalies/) and be somewhat more robust to temporary 'spikes' in the data that tend to always be happening somewhere in your metrics but often are not the most important type of anomaly (this is all covered in a lot more detail in the [deepdive tutorial](https://nbviewer.jupyter.org/github/netdata/community/blob/main/netdata-agent-api/netdata-pandas/anomalies_collector_deepdive.ipynb)).
-- You can see how long model training is taking by looking in the logs for the collector `grep 'anomalies' /var/log/netdata/error.log | grep 'training'` and you should see lines like `2020-12-01 22:02:14: python.d INFO: anomalies[local] : training complete in 2.81 seconds (runs_counter=2700, model=pca, train_n_secs=14400, models=26, n_fit_success=26, n_fit_fails=0, after=1606845731, before=1606860131).`.
- - This also gives counts of the number of models, if any, that failed to fit and so had to default back to the DefaultModel (which is currently [HBOS](https://pyod.readthedocs.io/en/latest/_modules/pyod/models/hbos.html)).
- - `after` and `before` here refer to the start and end of the training data used to train the models.
-- On a development n1-standard-2 (2 vCPUs, 7.5 GB memory) vm running Ubuntu 18.04 LTS and not doing any work some of the typical performance characteristics we saw from running this collector (with defaults) were:
- - A runtime (`netdata.runtime_anomalies`) of ~80ms when doing scoring and ~3 seconds when training or retraining the models.
- - Typically ~3%-3.5% additional cpu usage from scoring, jumping to ~60% for a couple of seconds during model training.
- - About ~150mb of ram (`apps.mem`) being continually used by the `python.d.plugin`.
-- If you activate this collector on a fresh node, it might take a little while to build up enough data to calculate a realistic and useful model.
-- Some models like `iforest` can be comparatively expensive (on same n1-standard-2 system above ~2s runtime during predict, ~40s training time, ~50% cpu on both train and predict) so if you would like to use it you might be advised to set a relatively high `update_every` maybe 10, 15 or 30 in `anomalies.conf`.
-- Setting a higher `train_every_n` and `update_every` is an easy way to devote less resources on the node to anomaly detection. Specifying less charts and a lower `train_n_secs` will also help reduce resources at the expense of covering less charts and maybe a more noisy model if you set `train_n_secs` to be too small for how your node tends to behave.
-- If you would like to enable this on a Raspberry Pi, then check out [this guide](https://github.com/netdata/netdata/blob/master/docs/guides/monitor/raspberry-pi-anomaly-detection.md) which will guide you through first installing LLVM.
-
-## Useful links and further reading
-
-- [PyOD documentation](https://pyod.readthedocs.io/en/latest/), [PyOD Github](https://github.com/yzhao062/pyod).
-- [Anomaly Detection](https://en.wikipedia.org/wiki/Anomaly_detection) wikipedia page.
-- [Anomaly Detection YouTube playlist](https://www.youtube.com/playlist?list=PL6Zhl9mK2r0KxA6rB87oi4kWzoqGd5vp0) maintained by [andrewm4894](https://github.com/andrewm4894/) from Netdata.
-- [awesome-TS-anomaly-detection](https://github.com/rob-med/awesome-TS-anomaly-detection) Github list of useful tools, libraries and resources.
-- [Mendeley public group](https://www.mendeley.com/community/interesting-anomaly-detection-papers/) with some interesting anomaly detection papers we have been reading.
-- Good [blog post](https://www.anodot.com/blog/what-is-anomaly-detection/) from Anodot on time series anomaly detection. Anodot also have some great whitepapers in this space too that some may find useful.
-- Novelty and outlier detection in the [scikit-learn documentation](https://scikit-learn.org/stable/modules/outlier_detection.html).
-
diff --git a/collectors/python.d.plugin/anomalies/metadata.yaml b/collectors/python.d.plugin/anomalies/metadata.yaml
deleted file mode 100644
index d138cf5dd..000000000
--- a/collectors/python.d.plugin/anomalies/metadata.yaml
+++ /dev/null
@@ -1,87 +0,0 @@
-# NOTE: this file is commented out as users are reccomended to use the
-# native anomaly detection capabilities on the agent instead.
-# meta:
-# plugin_name: python.d.plugin
-# module_name: anomalies
-# monitored_instance:
-# name: python.d anomalies
-# link: ""
-# categories: []
-# icon_filename: ""
-# related_resources:
-# integrations:
-# list: []
-# info_provided_to_referring_integrations:
-# description: ""
-# keywords: []
-# most_popular: false
-# overview:
-# data_collection:
-# metrics_description: ""
-# method_description: ""
-# supported_platforms:
-# include: []
-# exclude: []
-# multi_instance: true
-# additional_permissions:
-# description: ""
-# default_behavior:
-# auto_detection:
-# description: ""
-# limits:
-# description: ""
-# performance_impact:
-# description: ""
-# setup:
-# prerequisites:
-# list: []
-# configuration:
-# file:
-# name: ""
-# description: ""
-# options:
-# description: ""
-# folding:
-# title: ""
-# enabled: true
-# list: []
-# examples:
-# folding:
-# enabled: true
-# title: ""
-# list: []
-# troubleshooting:
-# problems:
-# list: []
-# alerts:
-# - name: anomalies_anomaly_probabilities
-# link: https://github.com/netdata/netdata/blob/master/health/health.d/anomalies.conf
-# metric: anomalies.probability
-# info: average anomaly probability over the last 2 minutes
-# - name: anomalies_anomaly_flags
-# link: https://github.com/netdata/netdata/blob/master/health/health.d/anomalies.conf
-# metric: anomalies.anomaly
-# info: number of anomalies in the last 2 minutes
-# metrics:
-# folding:
-# title: Metrics
-# enabled: false
-# description: ""
-# availability: []
-# scopes:
-# - name: global
-# description: ""
-# labels: []
-# metrics:
-# - name: anomalies.probability
-# description: Anomaly Probability
-# unit: "probability"
-# chart_type: line
-# dimensions:
-# - name: a dimension per probability
-# - name: anomalies.anomaly
-# description: Anomaly
-# unit: "count"
-# chart_type: stacked
-# dimensions:
-# - name: a dimension per anomaly
diff --git a/collectors/python.d.plugin/beanstalk/Makefile.inc b/collectors/python.d.plugin/beanstalk/Makefile.inc
deleted file mode 100644
index 4bbb7087d..000000000
--- a/collectors/python.d.plugin/beanstalk/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += beanstalk/beanstalk.chart.py
-dist_pythonconfig_DATA += beanstalk/beanstalk.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += beanstalk/README.md beanstalk/Makefile.inc
-
diff --git a/collectors/python.d.plugin/beanstalk/metadata.yaml b/collectors/python.d.plugin/beanstalk/metadata.yaml
deleted file mode 100644
index 7dff9cb3a..000000000
--- a/collectors/python.d.plugin/beanstalk/metadata.yaml
+++ /dev/null
@@ -1,263 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: beanstalk
- monitored_instance:
- name: Beanstalk
- link: "https://beanstalkd.github.io/"
- categories:
- - data-collection.message-brokers
- #- data-collection.task-queues
- icon_filename: "beanstalk.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - beanstalk
- - beanstalkd
- - message
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor Beanstalk metrics to enhance job queueing and processing efficiency. Track job rates, processing times, and queue lengths for better task management."
- method_description: "The collector uses the `beanstalkc` python module to connect to a `beanstalkd` service and gather metrics."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: "If no configuration is given, module will attempt to connect to beanstalkd on 127.0.0.1:11300 address."
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: "beanstalkc python module"
- description: The collector requires the `beanstalkc` python module to be installed.
- configuration:
- file:
- name: python.d/beanstalk.conf
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ""
- required: false
- - name: host
- description: IP or URL to a beanstalk service.
- default_value: "127.0.0.1"
- required: false
- - name: port
- description: Port to the IP or URL to a beanstalk service.
- default_value: "11300"
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Remote beanstalk server
- description: A basic remote beanstalk server
- folding:
- enabled: false
- config: |
- remote:
- name: 'beanstalk'
- host: '1.2.3.4'
- port: 11300
- - name: Multi-instance
- description: |
- > **Note**: When you define multiple jobs, their names must be unique.
-
- Collecting metrics from local and remote instances.
- config: |
- localhost:
- name: 'local_beanstalk'
- host: '127.0.0.1'
- port: 11300
-
- remote_job:
- name: 'remote_beanstalk'
- host: '192.0.2.1'
- port: 113000
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: beanstalk_server_buried_jobs
- link: https://github.com/netdata/netdata/blob/master/health/health.d/beanstalkd.conf
- metric: beanstalk.current_jobs
- info: number of buried jobs across all tubes. You need to manually kick them so they can be processed. Presence of buried jobs in a tube does not affect new jobs.
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: beanstalk.cpu_usage
- description: Cpu Usage
- unit: "cpu time"
- chart_type: area
- dimensions:
- - name: user
- - name: system
- - name: beanstalk.jobs_rate
- description: Jobs Rate
- unit: "jobs/s"
- chart_type: line
- dimensions:
- - name: total
- - name: timeouts
- - name: beanstalk.connections_rate
- description: Connections Rate
- unit: "connections/s"
- chart_type: area
- dimensions:
- - name: connections
- - name: beanstalk.commands_rate
- description: Commands Rate
- unit: "commands/s"
- chart_type: stacked
- dimensions:
- - name: put
- - name: peek
- - name: peek-ready
- - name: peek-delayed
- - name: peek-buried
- - name: reserve
- - name: use
- - name: watch
- - name: ignore
- - name: delete
- - name: bury
- - name: kick
- - name: stats
- - name: stats-job
- - name: stats-tube
- - name: list-tubes
- - name: list-tube-used
- - name: list-tubes-watched
- - name: pause-tube
- - name: beanstalk.connections_rate
- description: Current Tubes
- unit: "tubes"
- chart_type: area
- dimensions:
- - name: tubes
- - name: beanstalk.current_jobs
- description: Current Jobs
- unit: "jobs"
- chart_type: stacked
- dimensions:
- - name: urgent
- - name: ready
- - name: reserved
- - name: delayed
- - name: buried
- - name: beanstalk.current_connections
- description: Current Connections
- unit: "connections"
- chart_type: line
- dimensions:
- - name: written
- - name: producers
- - name: workers
- - name: waiting
- - name: beanstalk.binlog
- description: Binlog
- unit: "records/s"
- chart_type: line
- dimensions:
- - name: written
- - name: migrated
- - name: beanstalk.uptime
- description: seconds
- unit: "seconds"
- chart_type: line
- dimensions:
- - name: uptime
- - name: tube
- description: "Metrics related to Beanstalk tubes. Each tube produces its own set of the following metrics."
- labels: []
- metrics:
- - name: beanstalk.jobs_rate
- description: Jobs Rate
- unit: "jobs/s"
- chart_type: area
- dimensions:
- - name: jobs
- - name: beanstalk.jobs
- description: Jobs
- unit: "jobs"
- chart_type: stacked
- dimensions:
- - name: urgent
- - name: ready
- - name: reserved
- - name: delayed
- - name: buried
- - name: beanstalk.connections
- description: Connections
- unit: "connections"
- chart_type: stacked
- dimensions:
- - name: using
- - name: waiting
- - name: watching
- - name: beanstalk.commands
- description: Commands
- unit: "commands/s"
- chart_type: stacked
- dimensions:
- - name: deletes
- - name: pauses
- - name: beanstalk.pause
- description: Pause
- unit: "seconds"
- chart_type: stacked
- dimensions:
- - name: since
- - name: left
diff --git a/collectors/python.d.plugin/bind_rndc/Makefile.inc b/collectors/python.d.plugin/bind_rndc/Makefile.inc
deleted file mode 100644
index 72f391492..000000000
--- a/collectors/python.d.plugin/bind_rndc/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += bind_rndc/bind_rndc.chart.py
-dist_pythonconfig_DATA += bind_rndc/bind_rndc.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += bind_rndc/README.md bind_rndc/Makefile.inc
-
diff --git a/collectors/python.d.plugin/bind_rndc/metadata.yaml b/collectors/python.d.plugin/bind_rndc/metadata.yaml
deleted file mode 100644
index e3568e448..000000000
--- a/collectors/python.d.plugin/bind_rndc/metadata.yaml
+++ /dev/null
@@ -1,191 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: bind_rndc
- monitored_instance:
- name: ISC Bind (RNDC)
- link: "https://www.isc.org/bind/"
- categories:
- - data-collection.dns-and-dhcp-servers
- icon_filename: "isc.png"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - dns
- - bind
- - server
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor ISCBind (RNDC) performance for optimal DNS server operations. Monitor query rates, response times, and error rates to ensure reliable DNS service delivery."
- method_description: "This collector uses the `rndc` tool to dump (named.stats) statistics then read them to gather Bind Name Server summary performance metrics."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: "If no configuration is given, the collector will attempt to read named.stats file at `/var/log/bind/named.stats`"
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: "Minimum bind version and permissions"
- description: "Version of bind must be >=9.6 and the Netdata user must have permissions to run `rndc stats`"
- - title: "Setup log rotate for bind stats"
- description: |
- BIND appends logs at EVERY RUN. It is NOT RECOMMENDED to set `update_every` below 30 sec.
- It is STRONGLY RECOMMENDED to create a `bind-rndc.conf` file for logrotate.
-
- To set up BIND to dump stats do the following:
-
- 1. Add to 'named.conf.options' options {}:
- `statistics-file "/var/log/bind/named.stats";`
-
- 2. Create bind/ directory in /var/log:
- `cd /var/log/ && mkdir bind`
-
- 3. Change owner of directory to 'bind' user:
- `chown bind bind/`
-
- 4. RELOAD (NOT restart) BIND:
- `systemctl reload bind9.service`
-
- 5. Run as a root 'rndc stats' to dump (BIND will create named.stats in new directory)
-
- To allow Netdata to run 'rndc stats' change '/etc/bind/rndc.key' group to netdata:
- `chown :netdata rndc.key`
-
- Last, BUT NOT least, is to create bind-rndc.conf in logrotate.d/:
- ```
- /var/log/bind/named.stats {
-
- daily
- rotate 4
- compress
- delaycompress
- create 0644 bind bind
- missingok
- postrotate
- rndc reload > /dev/null
- endscript
- }
- ```
- To test your logrotate conf file run as root:
- `logrotate /etc/logrotate.d/bind-rndc -d (debug dry-run mode)`
- configuration:
- file:
- name: python.d/bind_rndc.conf
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ""
- required: false
- - name: named_stats_path
- description: Path to the named stats, after being dumped by `nrdc`
- default_value: "/var/log/bind/named.stats"
- required: false
- examples:
- folding:
- enabled: false
- title: "Config"
- list:
- - name: Local bind stats
- description: Define a local path to bind stats file
- config: |
- local:
- named_stats_path: '/var/log/bind/named.stats'
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: bind_rndc_stats_file_size
- link: https://github.com/netdata/netdata/blob/master/health/health.d/bind_rndc.conf
- metric: bind_rndc.stats_size
- info: BIND statistics-file size
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: bind_rndc.name_server_statistics
- description: Name Server Statistics
- unit: "stats"
- chart_type: line
- dimensions:
- - name: requests
- - name: rejected_queries
- - name: success
- - name: failure
- - name: responses
- - name: duplicate
- - name: recursion
- - name: nxrrset
- - name: nxdomain
- - name: non_auth_answer
- - name: auth_answer
- - name: dropped_queries
- - name: bind_rndc.incoming_queries
- description: Incoming queries
- unit: "queries"
- chart_type: line
- dimensions:
- - name: a dimension per incoming query type
- - name: bind_rndc.outgoing_queries
- description: Outgoing queries
- unit: "queries"
- chart_type: line
- dimensions:
- - name: a dimension per outgoing query type
- - name: bind_rndc.stats_size
- description: Named Stats File Size
- unit: "MiB"
- chart_type: line
- dimensions:
- - name: stats_size
diff --git a/collectors/python.d.plugin/boinc/Makefile.inc b/collectors/python.d.plugin/boinc/Makefile.inc
deleted file mode 100644
index 319e19cfe..000000000
--- a/collectors/python.d.plugin/boinc/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += boinc/boinc.chart.py
-dist_pythonconfig_DATA += boinc/boinc.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += boinc/README.md boinc/Makefile.inc
-
diff --git a/collectors/python.d.plugin/boinc/metadata.yaml b/collectors/python.d.plugin/boinc/metadata.yaml
deleted file mode 100644
index 33a67ac34..000000000
--- a/collectors/python.d.plugin/boinc/metadata.yaml
+++ /dev/null
@@ -1,198 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: boinc
- monitored_instance:
- name: BOINC
- link: "https://boinc.berkeley.edu/"
- categories:
- - data-collection.distributed-computing-systems
- icon_filename: "bolt.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - boinc
- - distributed
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This collector monitors task counts for the Berkeley Open Infrastructure Networking Computing (BOINC) distributed computing client."
- method_description: "It uses the same RPC interface that the BOINC monitoring GUI does."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: "By default, the module will try to auto-detect the password to the RPC interface by looking in `/var/lib/boinc` for this file (this is the location most Linux distributions use for a system-wide BOINC installation), so things may just work without needing configuration for a local system."
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: "Boinc RPC interface"
- description: BOINC requires use of a password to access it's RPC interface. You can find this password in the `gui_rpc_auth.cfg` file in your BOINC directory.
- configuration:
- file:
- name: python.d/boinc.conf
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ""
- required: false
- - name: hostname
- description: Define a hostname where boinc is running.
- default_value: "localhost"
- required: false
- - name: port
- description: The port of boinc RPC interface.
- default_value: ""
- required: false
- - name: password
- description: Provide a password to connect to a boinc RPC interface.
- default_value: ""
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Configuration of a remote boinc instance
- description: A basic JOB configuration for a remote boinc instance
- folding:
- enabled: false
- config: |
- remote:
- hostname: '1.2.3.4'
- port: 1234
- password: 'some-password'
- - name: Multi-instance
- description: |
- > **Note**: When you define multiple jobs, their names must be unique.
-
- Collecting metrics from local and remote instances.
- config: |
- localhost:
- name: 'local'
- host: '127.0.0.1'
- port: 1234
- password: 'some-password'
-
- remote_job:
- name: 'remote'
- host: '192.0.2.1'
- port: 1234
- password: some-other-password
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: boinc_total_tasks
- link: https://github.com/netdata/netdata/blob/master/health/health.d/boinc.conf
- metric: boinc.tasks
- info: average number of total tasks over the last 10 minutes
- os: "*"
- - name: boinc_active_tasks
- link: https://github.com/netdata/netdata/blob/master/health/health.d/boinc.conf
- metric: boinc.tasks
- info: average number of active tasks over the last 10 minutes
- os: "*"
- - name: boinc_compute_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/boinc.conf
- metric: boinc.states
- info: average number of compute errors over the last 10 minutes
- os: "*"
- - name: boinc_upload_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/boinc.conf
- metric: boinc.states
- info: average number of failed uploads over the last 10 minutes
- os: "*"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: boinc.tasks
- description: Overall Tasks
- unit: "tasks"
- chart_type: line
- dimensions:
- - name: Total
- - name: Active
- - name: boinc.states
- description: Tasks per State
- unit: "tasks"
- chart_type: line
- dimensions:
- - name: New
- - name: Downloading
- - name: Ready to Run
- - name: Compute Errors
- - name: Uploading
- - name: Uploaded
- - name: Aborted
- - name: Failed Uploads
- - name: boinc.sched
- description: Tasks per Scheduler State
- unit: "tasks"
- chart_type: line
- dimensions:
- - name: Uninitialized
- - name: Preempted
- - name: Scheduled
- - name: boinc.process
- description: Tasks per Process State
- unit: "tasks"
- chart_type: line
- dimensions:
- - name: Uninitialized
- - name: Executing
- - name: Suspended
- - name: Aborted
- - name: Quit
- - name: Copy Pending
diff --git a/collectors/python.d.plugin/ceph/Makefile.inc b/collectors/python.d.plugin/ceph/Makefile.inc
deleted file mode 100644
index 15b039ef6..000000000
--- a/collectors/python.d.plugin/ceph/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += ceph/ceph.chart.py
-dist_pythonconfig_DATA += ceph/ceph.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += ceph/README.md ceph/Makefile.inc
-
diff --git a/collectors/python.d.plugin/ceph/metadata.yaml b/collectors/python.d.plugin/ceph/metadata.yaml
deleted file mode 100644
index 0f06470b1..000000000
--- a/collectors/python.d.plugin/ceph/metadata.yaml
+++ /dev/null
@@ -1,223 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: ceph
- monitored_instance:
- name: Ceph
- link: 'https://ceph.io/'
- categories:
- - data-collection.storage-mount-points-and-filesystems
- icon_filename: 'ceph.svg'
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ''
- keywords:
- - ceph
- - storage
- most_popular: false
- overview:
- data_collection:
- metrics_description: 'This collector monitors Ceph metrics about Cluster statistics, OSD usage, latency and Pool statistics.'
- method_description: 'Uses the `rados` python module to connect to a Ceph cluster.'
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ''
- default_behavior:
- auto_detection:
- description: ''
- limits:
- description: ''
- performance_impact:
- description: ''
- setup:
- prerequisites:
- list:
- - title: '`rados` python module'
- description: 'Make sure the `rados` python module is installed'
- - title: 'Granting read permissions to ceph group from keyring file'
- description: 'Execute: `chmod 640 /etc/ceph/ceph.client.admin.keyring`'
- - title: 'Create a specific rados_id'
- description: 'You can optionally create a rados_id to use instead of admin'
- configuration:
- file:
- name: python.d/ceph.conf
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ''
- required: false
- - name: config_file
- description: Ceph config file
- default_value: ''
- required: true
- - name: keyring_file
- description: Ceph keyring file. netdata user must be added into ceph group and keyring file must be read group permission.
- default_value: ''
- required: true
- - name: rados_id
- description: A rados user id to use for connecting to the Ceph cluster.
- default_value: 'admin'
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Basic local Ceph cluster
- description: A basic configuration to connect to a local Ceph cluster.
- folding:
- enabled: false
- config: |
- local:
- config_file: '/etc/ceph/ceph.conf'
- keyring_file: '/etc/ceph/ceph.client.admin.keyring'
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: ceph_cluster_space_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ceph.conf
- metric: ceph.general_usage
- info: cluster disk space utilization
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: ceph.general_usage
- description: Ceph General Space
- unit: "KiB"
- chart_type: stacked
- dimensions:
- - name: avail
- - name: used
- - name: ceph.general_objects
- description: Ceph General Objects
- unit: "objects"
- chart_type: area
- dimensions:
- - name: cluster
- - name: ceph.general_bytes
- description: Ceph General Read/Write Data/s
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: ceph.general_operations
- description: Ceph General Read/Write Operations/s
- unit: "operations"
- chart_type: area
- dimensions:
- - name: read
- - name: write
- - name: ceph.general_latency
- description: Ceph General Apply/Commit latency
- unit: "milliseconds"
- chart_type: area
- dimensions:
- - name: apply
- - name: commit
- - name: ceph.pool_usage
- description: Ceph Pools
- unit: "KiB"
- chart_type: line
- dimensions:
- - name: a dimension per Ceph Pool
- - name: ceph.pool_objects
- description: Ceph Pools
- unit: "objects"
- chart_type: line
- dimensions:
- - name: a dimension per Ceph Pool
- - name: ceph.pool_read_bytes
- description: Ceph Read Pool Data/s
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: a dimension per Ceph Pool
- - name: ceph.pool_write_bytes
- description: Ceph Write Pool Data/s
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: a dimension per Ceph Pool
- - name: ceph.pool_read_operations
- description: Ceph Read Pool Operations/s
- unit: "operations"
- chart_type: area
- dimensions:
- - name: a dimension per Ceph Pool
- - name: ceph.pool_write_operations
- description: Ceph Write Pool Operations/s
- unit: "operations"
- chart_type: area
- dimensions:
- - name: a dimension per Ceph Pool
- - name: ceph.osd_usage
- description: Ceph OSDs
- unit: "KiB"
- chart_type: line
- dimensions:
- - name: a dimension per Ceph OSD
- - name: ceph.osd_size
- description: Ceph OSDs size
- unit: "KiB"
- chart_type: line
- dimensions:
- - name: a dimension per Ceph OSD
- - name: ceph.apply_latency
- description: Ceph OSDs apply latency
- unit: "milliseconds"
- chart_type: line
- dimensions:
- - name: a dimension per Ceph OSD
- - name: ceph.commit_latency
- description: Ceph OSDs commit latency
- unit: "milliseconds"
- chart_type: line
- dimensions:
- - name: a dimension per Ceph OSD
diff --git a/collectors/python.d.plugin/changefinder/Makefile.inc b/collectors/python.d.plugin/changefinder/Makefile.inc
deleted file mode 100644
index 01a92408b..000000000
--- a/collectors/python.d.plugin/changefinder/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += changefinder/changefinder.chart.py
-dist_pythonconfig_DATA += changefinder/changefinder.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += changefinder/README.md changefinder/Makefile.inc
-
diff --git a/collectors/python.d.plugin/dovecot/Makefile.inc b/collectors/python.d.plugin/dovecot/Makefile.inc
deleted file mode 100644
index fd7d13bbb..000000000
--- a/collectors/python.d.plugin/dovecot/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += dovecot/dovecot.chart.py
-dist_pythonconfig_DATA += dovecot/dovecot.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += dovecot/README.md dovecot/Makefile.inc
-
diff --git a/collectors/python.d.plugin/example/Makefile.inc b/collectors/python.d.plugin/example/Makefile.inc
deleted file mode 100644
index 1b027d5a7..000000000
--- a/collectors/python.d.plugin/example/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += example/example.chart.py
-dist_pythonconfig_DATA += example/example.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += example/README.md example/Makefile.inc
-
diff --git a/collectors/python.d.plugin/example/metadata.yaml b/collectors/python.d.plugin/example/metadata.yaml
deleted file mode 100644
index eae84d9e1..000000000
--- a/collectors/python.d.plugin/example/metadata.yaml
+++ /dev/null
@@ -1,138 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: example
- monitored_instance:
- name: Example collector
- link: https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/example/README.md
- categories:
- - data-collection.other
- icon_filename: ""
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - example
- - netdata
- - python
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- Example collector that generates some random numbers as metrics.
-
- If you want to write your own collector, read our [writing a new Python module](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/README.md#how-to-write-a-new-module) tutorial.
- method_description: |
- The `get_data()` function uses `random.randint()` to generate a random number which will be collected as a metric.
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: python.d/example.conf
- description: ""
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: Config options
- enabled: true
- list:
- - name: num_lines
- description: The number of lines to create.
- default_value: 4
- required: false
- - name: lower
- description: The lower bound of numbers to randomly sample from.
- default_value: 0
- required: false
- - name: upper
- description: The upper bound of numbers to randomly sample from.
- default_value: 100
- required: false
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 1
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ""
- required: false
- examples:
- folding:
- enabled: true
- title: Config
- list:
- - name: Basic
- folding:
- enabled: false
- description: A basic example configuration.
- config: |
- four_lines:
- name: "Four Lines"
- update_every: 1
- priority: 60000
- penalty: yes
- autodetection_retry: 0
- num_lines: 4
- lower: 0
- upper: 100
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: |
- These metrics refer to the entire monitored application.
- labels: []
- metrics:
- - name: example.random
- description: A random number
- unit: number
- chart_type: line
- dimensions:
- - name: random
diff --git a/collectors/python.d.plugin/exim/Makefile.inc b/collectors/python.d.plugin/exim/Makefile.inc
deleted file mode 100644
index 36ffa56d2..000000000
--- a/collectors/python.d.plugin/exim/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += exim/exim.chart.py
-dist_pythonconfig_DATA += exim/exim.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += exim/README.md exim/Makefile.inc
-
diff --git a/collectors/python.d.plugin/fail2ban/Makefile.inc b/collectors/python.d.plugin/fail2ban/Makefile.inc
deleted file mode 100644
index 31e117e53..000000000
--- a/collectors/python.d.plugin/fail2ban/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += fail2ban/fail2ban.chart.py
-dist_pythonconfig_DATA += fail2ban/fail2ban.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += fail2ban/README.md fail2ban/Makefile.inc
-
diff --git a/collectors/python.d.plugin/gearman/Makefile.inc b/collectors/python.d.plugin/gearman/Makefile.inc
deleted file mode 100644
index 275adf1c1..000000000
--- a/collectors/python.d.plugin/gearman/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += gearman/gearman.chart.py
-dist_pythonconfig_DATA += gearman/gearman.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += gearman/README.md gearman/Makefile.inc
-
diff --git a/collectors/python.d.plugin/gearman/metadata.yaml b/collectors/python.d.plugin/gearman/metadata.yaml
deleted file mode 100644
index f1760568e..000000000
--- a/collectors/python.d.plugin/gearman/metadata.yaml
+++ /dev/null
@@ -1,168 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: gearman
- monitored_instance:
- name: Gearman
- link: "http://gearman.org/"
- categories:
- - data-collection.distributed-computing-systems
- icon_filename: "gearman.png"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - gearman
- - gearman job server
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor Gearman metrics for proficient system task distribution. Track job counts, worker statuses, and queue lengths for effective distributed task management."
- method_description: "This collector connects to a Gearman instance via either TCP or unix socket."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: "When no configuration file is found, the collector tries to connect to TCP/IP socket: localhost:4730."
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: "Socket permissions"
- description: The gearman UNIX socket should have read permission for user netdata.
- configuration:
- file:
- name: python.d/gearman.conf
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ""
- required: false
- - name: host
- description: URL or IP where gearman is running.
- default_value: "localhost"
- required: false
- - name: port
- description: Port of URL or IP where gearman is running.
- default_value: "4730"
- required: false
- - name: tls
- description: Use tls to connect to gearman.
- default_value: "false"
- required: false
- - name: cert
- description: Provide a certificate file if needed to connect to a TLS gearman instance.
- default_value: ""
- required: false
- - name: key
- description: Provide a key file if needed to connect to a TLS gearman instance.
- default_value: ""
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Local gearman service
- description: A basic host and port gearman configuration for localhost.
- folding:
- enabled: false
- config: |
- localhost:
- name: 'local'
- host: 'localhost'
- port: 4730
- - name: Multi-instance
- description: |
- > **Note**: When you define multiple jobs, their names must be unique.
-
- Collecting metrics from local and remote instances.
- config: |
- localhost:
- name: 'local'
- host: 'localhost'
- port: 4730
-
- remote:
- name: 'remote'
- host: '192.0.2.1'
- port: 4730
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: gearman_workers_queued
- link: https://github.com/netdata/netdata/blob/master/health/health.d/gearman.conf
- metric: gearman.single_job
- info: average number of queued jobs over the last 10 minutes
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: gearman.total_jobs
- description: Total Jobs
- unit: "Jobs"
- chart_type: line
- dimensions:
- - name: Pending
- - name: Running
- - name: gearman job
- description: "Metrics related to Gearman jobs. Each job produces its own set of the following metrics."
- labels: []
- metrics:
- - name: gearman.single_job
- description: "{job_name}"
- unit: "Jobs"
- chart_type: stacked
- dimensions:
- - name: Pending
- - name: Idle
- - name: Runnning
diff --git a/collectors/python.d.plugin/go_expvar/Makefile.inc b/collectors/python.d.plugin/go_expvar/Makefile.inc
deleted file mode 100644
index 74f50d765..000000000
--- a/collectors/python.d.plugin/go_expvar/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += go_expvar/go_expvar.chart.py
-dist_pythonconfig_DATA += go_expvar/go_expvar.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += go_expvar/README.md go_expvar/Makefile.inc
-
diff --git a/collectors/python.d.plugin/go_expvar/metadata.yaml b/collectors/python.d.plugin/go_expvar/metadata.yaml
deleted file mode 100644
index 9419b024a..000000000
--- a/collectors/python.d.plugin/go_expvar/metadata.yaml
+++ /dev/null
@@ -1,329 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: go_expvar
- monitored_instance:
- name: Go applications (EXPVAR)
- link: "https://pkg.go.dev/expvar"
- categories:
- - data-collection.apm
- icon_filename: "go.png"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - go
- - expvar
- - application
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This collector monitors Go applications that expose their metrics with the use of the `expvar` package from the Go standard library. It produces charts for Go runtime memory statistics and optionally any number of custom charts."
- method_description: "It connects via http to gather the metrics exposed via the `expvar` package."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: "Enable the go_expvar collector"
- description: |
- The `go_expvar` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`, to edit the `python.d.conf` file.
-
- ```bash
- cd /etc/netdata # Replace this path with your Netdata config directory, if different
- sudo ./edit-config python.d.conf
- ```
-
- Change the value of the `go_expvar` setting to `yes`. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
- - title: "Sample `expvar` usage in a Go application"
- description: |
- The `expvar` package exposes metrics over HTTP and is very easy to use.
- Consider this minimal sample below:
-
- ```go
- package main
-
- import (
- _ "expvar"
- "net/http"
- )
-
- func main() {
- http.ListenAndServe("127.0.0.1:8080", nil)
- }
- ```
-
- When imported this way, the `expvar` package registers a HTTP handler at `/debug/vars` that
- exposes Go runtime's memory statistics in JSON format. You can inspect the output by opening
- the URL in your browser (or by using `wget` or `curl`).
-
- Sample output:
-
- ```json
- {
- "cmdline": ["./expvar-demo-binary"],
- "memstats": {"Alloc":630856,"TotalAlloc":630856,"Sys":3346432,"Lookups":27, <omitted for brevity>}
- }
- ```
-
- You can of course expose and monitor your own variables as well.
- Here is a sample Go application that exposes a few custom variables:
-
- ```go
- package main
-
- import (
- "expvar"
- "net/http"
- "runtime"
- "time"
- )
-
- func main() {
-
- tick := time.NewTicker(1 * time.Second)
- num_go := expvar.NewInt("runtime.goroutines")
- counters := expvar.NewMap("counters")
- counters.Set("cnt1", new(expvar.Int))
- counters.Set("cnt2", new(expvar.Float))
-
- go http.ListenAndServe(":8080", nil)
-
- for {
- select {
- case <- tick.C:
- num_go.Set(int64(runtime.NumGoroutine()))
- counters.Add("cnt1", 1)
- counters.AddFloat("cnt2", 1.452)
- }
- }
- }
- ```
-
- Apart from the runtime memory stats, this application publishes two counters and the
- number of currently running Goroutines and updates these stats every second.
- configuration:
- file:
- name: python.d/go_expvar.conf
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified. Each JOB can be used to monitor a different Go application.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ""
- required: false
- - name: url
- description: the URL and port of the expvar endpoint. Please include the whole path of the endpoint, as the expvar handler can be installed in a non-standard location.
- default_value: ""
- required: true
- - name: user
- description: If the URL is password protected, this is the username to use.
- default_value: ""
- required: false
- - name: pass
- description: If the URL is password protected, this is the password to use.
- default_value: ""
- required: false
- - name: collect_memstats
- description: Enables charts for Go runtime's memory statistics.
- default_value: ""
- required: false
- - name: extra_charts
- description: Defines extra data/charts to monitor, please see the example below.
- default_value: ""
- required: false
- examples:
- folding:
- enabled: false
- title: "Config"
- list:
- - name: Monitor a Go app1 application
- description: |
- The example below sets a configuration for a Go application, called `app1`. Besides the `memstats`, the application also exposes two counters and the number of currently running Goroutines and updates these stats every second.
-
- The `go_expvar` collector can monitor these as well with the use of the `extra_charts` configuration variable.
-
- The `extra_charts` variable is a YaML list of Netdata chart definitions.
- Each chart definition has the following keys:
-
- ```
- id: Netdata chart ID
- options: a key-value mapping of chart options
- lines: a list of line definitions
- ```
-
- **Note: please do not use dots in the chart or line ID field.
- See [this issue](https://github.com/netdata/netdata/pull/1902#issuecomment-284494195) for explanation.**
-
- Please see these two links to the official Netdata documentation for more information about the values:
-
- - [External plugins - charts](https://github.com/netdata/netdata/blob/master/collectors/plugins.d/README.md#chart)
- - [Chart variables](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/README.md#global-variables-order-and-chart)
-
- **Line definitions**
-
- Each chart can define multiple lines (dimensions).
- A line definition is a key-value mapping of line options.
- Each line can have the following options:
-
- ```
- # mandatory
- expvar_key: the name of the expvar as present in the JSON output of /debug/vars endpoint
- expvar_type: value type; supported are "float" or "int"
- id: the id of this line/dimension in Netdata
-
- # optional - Netdata defaults are used if these options are not defined
- name: ''
- algorithm: absolute
- multiplier: 1
- divisor: 100 if expvar_type == float, 1 if expvar_type == int
- hidden: False
- ```
-
- Please see the following link for more information about the options and their default values:
- [External plugins - dimensions](https://github.com/netdata/netdata/blob/master/collectors/plugins.d/README.md#dimension)
-
- Apart from top-level expvars, this plugin can also parse expvars stored in a multi-level map;
- All dicts in the resulting JSON document are then flattened to one level.
- Expvar names are joined together with '.' when flattening.
-
- Example:
-
- ```
- {
- "counters": {"cnt1": 1042, "cnt2": 1512.9839999999983},
- "runtime.goroutines": 5
- }
- ```
-
- In the above case, the exported variables will be available under `runtime.goroutines`,
- `counters.cnt1` and `counters.cnt2` expvar_keys. If the flattening results in a key collision,
- the first defined key wins and all subsequent keys with the same name are ignored.
- config: |
- app1:
- name : 'app1'
- url : 'http://127.0.0.1:8080/debug/vars'
- collect_memstats: true
- extra_charts:
- - id: "runtime_goroutines"
- options:
- name: num_goroutines
- title: "runtime: number of goroutines"
- units: goroutines
- family: runtime
- context: expvar.runtime.goroutines
- chart_type: line
- lines:
- - {expvar_key: 'runtime.goroutines', expvar_type: int, id: runtime_goroutines}
- - id: "foo_counters"
- options:
- name: counters
- title: "some random counters"
- units: awesomeness
- family: counters
- context: expvar.foo.counters
- chart_type: line
- lines:
- - {expvar_key: 'counters.cnt1', expvar_type: int, id: counters_cnt1}
- - {expvar_key: 'counters.cnt2', expvar_type: float, id: counters_cnt2}
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: expvar.memstats.heap
- description: "memory: size of heap memory structures"
- unit: "KiB"
- chart_type: line
- dimensions:
- - name: alloc
- - name: inuse
- - name: expvar.memstats.stack
- description: "memory: size of stack memory structures"
- unit: "KiB"
- chart_type: line
- dimensions:
- - name: inuse
- - name: expvar.memstats.mspan
- description: "memory: size of mspan memory structures"
- unit: "KiB"
- chart_type: line
- dimensions:
- - name: inuse
- - name: expvar.memstats.mcache
- description: "memory: size of mcache memory structures"
- unit: "KiB"
- chart_type: line
- dimensions:
- - name: inuse
- - name: expvar.memstats.live_objects
- description: "memory: number of live objects"
- unit: "objects"
- chart_type: line
- dimensions:
- - name: live
- - name: expvar.memstats.sys
- description: "memory: size of reserved virtual address space"
- unit: "KiB"
- chart_type: line
- dimensions:
- - name: sys
- - name: expvar.memstats.gc_pauses
- description: "memory: average duration of GC pauses"
- unit: "ns"
- chart_type: line
- dimensions:
- - name: avg
diff --git a/collectors/python.d.plugin/haproxy/Makefile.inc b/collectors/python.d.plugin/haproxy/Makefile.inc
deleted file mode 100644
index ad24deaa0..000000000
--- a/collectors/python.d.plugin/haproxy/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += haproxy/haproxy.chart.py
-dist_pythonconfig_DATA += haproxy/haproxy.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += haproxy/README.md haproxy/Makefile.inc
-
diff --git a/collectors/python.d.plugin/haproxy/README.md b/collectors/python.d.plugin/haproxy/README.md
deleted file mode 100644
index 2fa203f60..000000000
--- a/collectors/python.d.plugin/haproxy/README.md
+++ /dev/null
@@ -1,90 +0,0 @@
-<!--
-title: "HAProxy monitoring with Netdata"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/haproxy/README.md"
-sidebar_label: "haproxy-python.d.plugin"
-learn_status: "Published"
-learn_topic_type: "References"
-learn_rel_path: "Integrations/Monitor/Webapps"
--->
-
-# HAProxy collector
-
-Monitors frontend and backend metrics such as bytes in, bytes out, sessions current, sessions in queue current.
-And health metrics such as backend servers status (server check should be used).
-
-Plugin can obtain data from URL or Unix socket.
-
-Requirement:
-
-- Socket must be readable and writable by the `netdata` user.
-- URL must have `stats uri <path>` present in the haproxy config, otherwise you will get HTTP 503 in the haproxy logs.
-
-It produces:
-
-1. **Frontend** family charts
-
- - Kilobytes in/s
- - Kilobytes out/s
- - Sessions current
- - Sessions in queue current
-
-2. **Backend** family charts
-
- - Kilobytes in/s
- - Kilobytes out/s
- - Sessions current
- - Sessions in queue current
-
-3. **Health** chart
-
- - number of failed servers for every backend (in DOWN state)
-
-## Configuration
-
-Edit the `python.d/haproxy.conf` configuration file using `edit-config` from the Netdata [config
-directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`.
-
-```bash
-cd /etc/netdata # Replace this path with your Netdata config directory, if different
-sudo ./edit-config python.d/haproxy.conf
-```
-
-Sample:
-
-```yaml
-via_url:
- user: 'username' # ONLY IF stats auth is used
- pass: 'password' # # ONLY IF stats auth is used
- url: 'http://ip.address:port/url;csv;norefresh'
-```
-
-OR
-
-```yaml
-via_socket:
- socket: 'path/to/haproxy/sock'
-```
-
-If no configuration is given, module will fail to run.
-
-
-### Troubleshooting
-
-To troubleshoot issues with the `haproxy` module, run the `python.d.plugin` with the debug option enabled. The
-output will give you the output of the data collection job or error messages on why the collector isn't working.
-
-First, navigate to your plugins directory, usually they are located under `/usr/libexec/netdata/plugins.d/`. If that's
-not the case on your system, open `netdata.conf` and look for the setting `plugins directory`. Once you're in the
-plugin's directory, switch to the `netdata` user.
-
-```bash
-cd /usr/libexec/netdata/plugins.d/
-sudo su -s /bin/bash netdata
-```
-
-Now you can manually run the `haproxy` module in debug mode:
-
-```bash
-./python.d.plugin haproxy debug trace
-```
-
diff --git a/collectors/python.d.plugin/haproxy/metadata.yaml b/collectors/python.d.plugin/haproxy/metadata.yaml
deleted file mode 100644
index 82ab37d26..000000000
--- a/collectors/python.d.plugin/haproxy/metadata.yaml
+++ /dev/null
@@ -1,322 +0,0 @@
-# This collector will not appear in documentation, as the go version is preferred,
-# https://github.com/netdata/go.d.plugin/blob/master/modules/haproxy/README.md
-#
-#
-# meta:
-# plugin_name: python.d.plugin
-# module_name: haproxy
-# monitored_instance:
-# name: HAProxy
-# link: 'https://www.haproxy.org/'
-# categories:
-# - data-collection.web-servers-and-web-proxies
-# icon_filename: 'haproxy.png'
-# related_resources:
-# integrations:
-# list: []
-# info_provided_to_referring_integrations:
-# description: ''
-# keywords:
-# - haproxy
-# - tcp
-# - balancer
-# most_popular: false
-# overview:
-# data_collection:
-# metrics_description: 'This collector monitors HAProxy metrics about frontend servers, backend servers, responses and more.'
-# method_description: 'It connects to the HAProxy instance via URL or UNIX socket.'
-# supported_platforms:
-# include: []
-# exclude: []
-# multi_instance: true
-# additional_permissions:
-# description: ''
-# default_behavior:
-# auto_detection:
-# description: ''
-# limits:
-# description: ''
-# performance_impact:
-# description: ''
-# setup:
-# prerequisites:
-# list:
-# - title: 'HAProxy setup for socket'
-# description: 'Socket must be readable and writable by the netdata user.'
-# - title: 'HAProxy setup for URL'
-# description: 'URL must have `stats uri <path>` present in the haproxy config, otherwise you will get HTTP 503 in the haproxy logs.'
-# configuration:
-# file:
-# name: python.d/haproxy.conf
-# options:
-# description: |
-# There are 2 sections:
-
-# * Global variables
-# * One or more JOBS that can define multiple different instances to monitor.
-
-# The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
-# Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
-# Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
-# folding:
-# title: "Config options"
-# enabled: true
-# list:
-# - name: update_every
-# description: Sets the default data collection frequency.
-# default_value: 5
-# required: false
-# - name: priority
-# description: Controls the order of charts at the netdata dashboard.
-# default_value: 60000
-# required: false
-# - name: autodetection_retry
-# description: Sets the job re-check interval in seconds.
-# default_value: 0
-# required: false
-# - name: penalty
-# description: Indicates whether to apply penalty to update_every in case of failures.
-# default_value: yes
-# required: false
-# - name: name
-# description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
-# default_value: ''
-# required: false
-# - name: user
-# description: Username if stats auth is used.
-# default_value: ''
-# required: false
-# - name: pass
-# description: Password if stats auth is used.
-# default_value: ''
-# required: false
-# - name: url
-# description: URL to the haproxy_stats endpoint. Also make sure the parameters `csv` and `norefresh` are provided.
-# default_value: ''
-# required: false
-# - name: socket
-# description: Unix socket path to the haproxy sock file.
-# default_value: ''
-# required: false
-# examples:
-# folding:
-# enabled: true
-# title: "Config"
-# list:
-# - name: URL method
-# description: Use a URL to specify the endpoint to check for haproxy statistics.
-# config: |
-# via_url:
-# user: 'username' # ONLY IF stats auth is used
-# pass: 'password' # # ONLY IF stats auth is used
-# url: 'http://ip.address:port/url;csv;norefresh'
-# - name: Local socket
-# description: Use a local socket to check for haproxy statistics.
-# config: |
-# via_socket:
-# socket: 'path/to/haproxy/sock'
-# troubleshooting:
-# problems:
-# list: []
-# alerts:
-# - name: haproxy_backend_server_status
-# link: https://github.com/netdata/netdata/blob/master/health/health.d/haproxy.conf
-# metric: haproxy_hs.down
-# info: average number of failed haproxy backend servers over the last 10 seconds
-# - name: haproxy_backend_status
-# link: https://github.com/netdata/netdata/blob/master/health/health.d/haproxy.conf
-# metric: haproxy_hb.down
-# info: average number of failed haproxy backends over the last 10 seconds
-# metrics:
-# folding:
-# title: Metrics
-# enabled: false
-# description: ""
-# availability: []
-# scopes:
-# - name: global
-# description: 'These metrics refer to the entire monitored application.'
-# labels: []
-# metrics:
-# - name: haproxy_f.bin
-# description: Kilobytes In
-# unit: "KiB/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_f.bout
-# description: Kilobytes Out
-# unit: "KiB/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_f.scur
-# description: Sessions Active
-# unit: "sessions"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_f.qcur
-# description: Session In Queue
-# unit: "sessions"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_f.hrsp_1xx
-# description: HTTP responses with 1xx code
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_f.hrsp_2xx
-# description: HTTP responses with 2xx code
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_f.hrsp_3xx
-# description: HTTP responses with 3xx code
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_f.hrsp_4xx
-# description: HTTP responses with 4xx code
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_f.hrsp_5xx
-# description: HTTP responses with 5xx code
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_f.hrsp_other
-# description: HTTP responses with other codes (protocol error)
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_f.hrsp_total
-# description: HTTP responses
-# unit: "responses"
-# chart_type: line
-# dimensions:
-# - name: a dimension per frontend server
-# - name: haproxy_b.bin
-# description: Kilobytes In
-# unit: "KiB/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.bout
-# description: Kilobytes Out
-# unit: "KiB/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.scur
-# description: Sessions Active
-# unit: "sessions"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.qcur
-# description: Sessions In Queue
-# unit: "sessions"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.hrsp_1xx
-# description: HTTP responses with 1xx code
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.hrsp_2xx
-# description: HTTP responses with 2xx code
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.hrsp_3xx
-# description: HTTP responses with 3xx code
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.hrsp_4xx
-# description: HTTP responses with 4xx code
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.hrsp_5xx
-# description: HTTP responses with 5xx code
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.hrsp_other
-# description: HTTP responses with other codes (protocol error)
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.hrsp_total
-# description: HTTP responses (total)
-# unit: "responses/s"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.qtime
-# description: The average queue time over the 1024 last requests
-# unit: "milliseconds"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.ctime
-# description: The average connect time over the 1024 last requests
-# unit: "milliseconds"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.rtime
-# description: The average response time over the 1024 last requests
-# unit: "milliseconds"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_b.ttime
-# description: The average total session time over the 1024 last requests
-# unit: "milliseconds"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_hs.down
-# description: Backend Servers In DOWN State
-# unit: "failed servers"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_hs.up
-# description: Backend Servers In UP State
-# unit: "health servers"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy_hb.down
-# description: Is Backend Failed?
-# unit: "boolean"
-# chart_type: line
-# dimensions:
-# - name: a dimension per backend server
-# - name: haproxy.idle
-# description: The Ratio Of Polling Time Vs Total Time
-# unit: "percentage"
-# chart_type: line
-# dimensions:
-# - name: idle
diff --git a/collectors/python.d.plugin/hddtemp/Makefile.inc b/collectors/python.d.plugin/hddtemp/Makefile.inc
deleted file mode 100644
index 22852b646..000000000
--- a/collectors/python.d.plugin/hddtemp/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += hddtemp/hddtemp.chart.py
-dist_pythonconfig_DATA += hddtemp/hddtemp.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += hddtemp/README.md hddtemp/Makefile.inc
-
diff --git a/collectors/python.d.plugin/hpssa/Makefile.inc b/collectors/python.d.plugin/hpssa/Makefile.inc
deleted file mode 100644
index 1c04aa49c..000000000
--- a/collectors/python.d.plugin/hpssa/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += hpssa/hpssa.chart.py
-dist_pythonconfig_DATA += hpssa/hpssa.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += hpssa/README.md hpssa/Makefile.inc
-
diff --git a/collectors/python.d.plugin/hpssa/metadata.yaml b/collectors/python.d.plugin/hpssa/metadata.yaml
deleted file mode 100644
index 7871cc276..000000000
--- a/collectors/python.d.plugin/hpssa/metadata.yaml
+++ /dev/null
@@ -1,185 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: hpssa
- monitored_instance:
- name: HP Smart Storage Arrays
- link: 'https://buy.hpe.com/us/en/software/server-management-software/server-management-software/smart-array-management-software/hpe-smart-storage-administrator/p/5409020'
- categories:
- - data-collection.storage-mount-points-and-filesystems
- icon_filename: 'hp.svg'
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ''
- keywords:
- - storage
- - hp
- - hpssa
- - array
- most_popular: false
- overview:
- data_collection:
- metrics_description: 'This collector monitors HP Smart Storage Arrays metrics about operational statuses and temperatures.'
- method_description: 'It uses the command line tool `ssacli`. The exact command used is `sudo -n ssacli ctrl all show config detail`'
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: ''
- default_behavior:
- auto_detection:
- description: 'If no configuration is provided, the collector will try to execute the `ssacli` binary.'
- limits:
- description: ''
- performance_impact:
- description: ''
- setup:
- prerequisites:
- list:
- - title: 'Enable the hpssa collector'
- description: |
- The `hpssa` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`, to edit the `python.d.conf` file.
-
- ```bash
- cd /etc/netdata # Replace this path with your Netdata config directory, if different
- sudo ./edit-config python.d.conf
- ```
-
- Change the value of the `hpssa` setting to `yes`. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
- - title: 'Allow user netdata to execute `ssacli` as root.'
- description: |
- This module uses `ssacli`, which can only be executed by root. It uses `sudo` and assumes that it is configured such that the `netdata` user can execute `ssacli` as root without a password.
-
- - Add to your `/etc/sudoers` file:
-
- `which ssacli` shows the full path to the binary.
-
- ```bash
- netdata ALL=(root) NOPASSWD: /path/to/ssacli
- ```
-
- - Reset Netdata's systemd
- unit [CapabilityBoundingSet](https://www.freedesktop.org/software/systemd/man/systemd.exec.html#Capabilities) (Linux
- distributions with systemd)
-
- The default CapabilityBoundingSet doesn't allow using `sudo`, and is quite strict in general. Resetting is not optimal, but a next-best solution given the inability to execute `ssacli` using `sudo`.
-
- As the `root` user, do the following:
-
- ```cmd
- mkdir /etc/systemd/system/netdata.service.d
- echo -e '[Service]\nCapabilityBoundingSet=~' | tee /etc/systemd/system/netdata.service.d/unset-capability-bounding-set.conf
- systemctl daemon-reload
- systemctl restart netdata.service
- ```
- configuration:
- file:
- name: python.d/hpssa.conf
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ''
- required: false
- - name: ssacli_path
- description: Path to the `ssacli` command line utility. Configure this if `ssacli` is not in the $PATH
- default_value: ''
- required: false
- - name: use_sudo
- description: Whether or not to use `sudo` to execute `ssacli`
- default_value: 'True'
- required: false
- examples:
- folding:
- enabled: false
- title: "Config"
- list:
- - name: Local simple config
- description: A basic configuration, specyfing the path to `ssacli`
- folding:
- enabled: false
- config: |
- local:
- ssacli_path: /usr/sbin/ssacli
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: hpssa.ctrl_status
- description: Status 1 is OK, Status 0 is not OK
- unit: "Status"
- chart_type: line
- dimensions:
- - name: ctrl_{adapter slot}_status
- - name: cache_{adapter slot}_status
- - name: battery_{adapter slot}_status per adapter
- - name: hpssa.ctrl_temperature
- description: Temperature
- unit: "Celsius"
- chart_type: line
- dimensions:
- - name: ctrl_{adapter slot}_temperature
- - name: cache_{adapter slot}_temperature per adapter
- - name: hpssa.ld_status
- description: Status 1 is OK, Status 0 is not OK
- unit: "Status"
- chart_type: line
- dimensions:
- - name: a dimension per logical drive
- - name: hpssa.pd_status
- description: Status 1 is OK, Status 0 is not OK
- unit: "Status"
- chart_type: line
- dimensions:
- - name: a dimension per physical drive
- - name: hpssa.pd_temperature
- description: Temperature
- unit: "Celsius"
- chart_type: line
- dimensions:
- - name: a dimension per physical drive
diff --git a/collectors/python.d.plugin/icecast/Makefile.inc b/collectors/python.d.plugin/icecast/Makefile.inc
deleted file mode 100644
index cb7c6fa0e..000000000
--- a/collectors/python.d.plugin/icecast/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += icecast/icecast.chart.py
-dist_pythonconfig_DATA += icecast/icecast.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += icecast/README.md icecast/Makefile.inc
-
diff --git a/collectors/python.d.plugin/ipfs/Makefile.inc b/collectors/python.d.plugin/ipfs/Makefile.inc
deleted file mode 100644
index 68458cb38..000000000
--- a/collectors/python.d.plugin/ipfs/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += ipfs/ipfs.chart.py
-dist_pythonconfig_DATA += ipfs/ipfs.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += ipfs/README.md ipfs/Makefile.inc
-
diff --git a/collectors/python.d.plugin/ipfs/metadata.yaml b/collectors/python.d.plugin/ipfs/metadata.yaml
deleted file mode 100644
index dbc421c90..000000000
--- a/collectors/python.d.plugin/ipfs/metadata.yaml
+++ /dev/null
@@ -1,172 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: ipfs
- monitored_instance:
- name: IPFS
- link: "https://ipfs.tech/"
- categories:
- - data-collection.storage-mount-points-and-filesystems
- icon_filename: "ipfs.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This collector monitors IPFS server metrics about its quality and performance."
- method_description: "It connects to an http endpoint of the IPFS server to collect the metrics"
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: "If the endpoint is accessible by the Agent, netdata will autodetect it"
- limits:
- description: |
- Calls to the following endpoints are disabled due to IPFS bugs:
-
- /api/v0/stats/repo (https://github.com/ipfs/go-ipfs/issues/3874)
- /api/v0/pin/ls (https://github.com/ipfs/go-ipfs/issues/7528)
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "python.d/ipfs.conf"
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: ""
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: The JOB's name as it will appear at the dashboard (by default is the job_name)
- default_value: job_name
- required: false
- - name: url
- description: URL to the IPFS API
- default_value: no
- required: true
- - name: repoapi
- description: Collect repo metrics.
- default_value: no
- required: false
- - name: pinapi
- description: Set status of IPFS pinned object polling.
- default_value: no
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Basic (default out-of-the-box)
- description: A basic example configuration, one job will run at a time. Autodetect mechanism uses it by default.
- folding:
- enabled: false
- config: |
- localhost:
- name: 'local'
- url: 'http://localhost:5001'
- repoapi: no
- pinapi: no
- - name: Multi-instance
- description: |
- > **Note**: When you define multiple jobs, their names must be unique.
-
- Collecting metrics from local and remote instances.
- config: |
- localhost:
- name: 'local'
- url: 'http://localhost:5001'
- repoapi: no
- pinapi: no
-
- remote_host:
- name: 'remote'
- url: 'http://192.0.2.1:5001'
- repoapi: no
- pinapi: no
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: ipfs_datastore_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/ipfs.conf
- metric: ipfs.repo_size
- info: IPFS datastore utilization
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: ipfs.bandwidth
- description: IPFS Bandwidth
- unit: "kilobits/s"
- chart_type: line
- dimensions:
- - name: in
- - name: out
- - name: ipfs.peers
- description: IPFS Peers
- unit: "peers"
- chart_type: line
- dimensions:
- - name: peers
- - name: ipfs.repo_size
- description: IPFS Repo Size
- unit: "GiB"
- chart_type: area
- dimensions:
- - name: avail
- - name: size
- - name: ipfs.repo_objects
- description: IPFS Repo Objects
- unit: "objects"
- chart_type: line
- dimensions:
- - name: objects
- - name: pinned
- - name: recursive_pins
diff --git a/collectors/python.d.plugin/litespeed/Makefile.inc b/collectors/python.d.plugin/litespeed/Makefile.inc
deleted file mode 100644
index 5dd645020..000000000
--- a/collectors/python.d.plugin/litespeed/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += litespeed/litespeed.chart.py
-dist_pythonconfig_DATA += litespeed/litespeed.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += litespeed/README.md litespeed/Makefile.inc
-
diff --git a/collectors/python.d.plugin/megacli/Makefile.inc b/collectors/python.d.plugin/megacli/Makefile.inc
deleted file mode 100644
index 83680d723..000000000
--- a/collectors/python.d.plugin/megacli/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += megacli/megacli.chart.py
-dist_pythonconfig_DATA += megacli/megacli.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += megacli/README.md megacli/Makefile.inc
-
diff --git a/collectors/python.d.plugin/megacli/metadata.yaml b/collectors/python.d.plugin/megacli/metadata.yaml
deleted file mode 100644
index 4a2ba43ee..000000000
--- a/collectors/python.d.plugin/megacli/metadata.yaml
+++ /dev/null
@@ -1,193 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: megacli
- monitored_instance:
- name: MegaCLI
- link: "https://wikitech.wikimedia.org/wiki/MegaCli"
- categories:
- - data-collection.storage-mount-points-and-filesystems
- icon_filename: "hard-drive.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - storage
- - raid-controller
- - manage-disks
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Examine MegaCLI metrics with Netdata for insights into RAID controller performance. Improve your RAID controller efficiency with real-time MegaCLI metrics."
- method_description: |
- Collects adapter, physical drives and battery stats using megacli command-line tool
-
- Executed commands:
-
- - `sudo -n megacli -LDPDInfo -aAll`
- - `sudo -n megacli -AdpBbuCmd -a0`
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: "The module uses megacli, which can only be executed by root. It uses sudo and assumes that it is configured such that the netdata user can execute megacli as root without a password."
- default_behavior:
- auto_detection:
- description: "After all the permissions are satisfied, netdata should be to execute commands via the megacli command line utility"
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: Grant permissions for netdata, to run megacli as sudoer
- description: |
- The module uses megacli, which can only be executed by root. It uses sudo and assumes that it is configured such that the netdata user can execute megacli as root without a password.
-
- Add to your /etc/sudoers file:
- which megacli shows the full path to the binary.
-
- ```bash
- netdata ALL=(root) NOPASSWD: /path/to/megacli
- ```
- - title: "Reset Netdata's systemd unit CapabilityBoundingSet (Linux distributions with systemd)"
- description: |
- The default CapabilityBoundingSet doesn't allow using sudo, and is quite strict in general. Resetting is not optimal, but a next-best solution given the inability to execute arcconf using sudo.
-
- As root user, do the following:
-
- ```bash
- mkdir /etc/systemd/system/netdata.service.d
- echo -e '[Service]\nCapabilityBoundingSet=~' | tee /etc/systemd/system/netdata.service.d/unset-capability-bounding-set.conf
- systemctl daemon-reload
- systemctl restart netdata.service
- ```
- configuration:
- file:
- name: "python.d/megacli.conf"
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: do_battery
- description: default is no. Battery stats (adds additional call to megacli `megacli -AdpBbuCmd -a0`).
- default_value: no
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Basic
- folding:
- enabled: false
- description: A basic example configuration per job
- config: |
- job_name:
- name: myname
- update_every: 1
- priority: 60000
- penalty: yes
- autodetection_retry: 0
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: megacli_adapter_state
- link: https://github.com/netdata/netdata/blob/master/health/health.d/megacli.conf
- metric: megacli.adapter_degraded
- info: "adapter is in the degraded state (0: false, 1: true)"
- - name: megacli_pd_media_errors
- link: https://github.com/netdata/netdata/blob/master/health/health.d/megacli.conf
- metric: megacli.pd_media_error
- info: number of physical drive media errors
- - name: megacli_pd_predictive_failures
- link: https://github.com/netdata/netdata/blob/master/health/health.d/megacli.conf
- metric: megacli.pd_predictive_failure
- info: number of physical drive predictive failures
- - name: megacli_bbu_relative_charge
- link: https://github.com/netdata/netdata/blob/master/health/health.d/megacli.conf
- metric: megacli.bbu_relative_charge
- info: average battery backup unit (BBU) relative state of charge over the last 10 seconds
- - name: megacli_bbu_cycle_count
- link: https://github.com/netdata/netdata/blob/master/health/health.d/megacli.conf
- metric: megacli.bbu_cycle_count
- info: average battery backup unit (BBU) charge cycles count over the last 10 seconds
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: megacli.adapter_degraded
- description: Adapter State
- unit: "is degraded"
- chart_type: line
- dimensions:
- - name: a dimension per adapter
- - name: megacli.pd_media_error
- description: Physical Drives Media Errors
- unit: "errors/s"
- chart_type: line
- dimensions:
- - name: a dimension per physical drive
- - name: megacli.pd_predictive_failure
- description: Physical Drives Predictive Failures
- unit: "failures/s"
- chart_type: line
- dimensions:
- - name: a dimension per physical drive
- - name: battery
- description: "Metrics related to Battery Backup Units, each BBU provides its own set of the following metrics."
- labels: []
- metrics:
- - name: megacli.bbu_relative_charge
- description: Relative State of Charge
- unit: "percentage"
- chart_type: line
- dimensions:
- - name: adapter {battery id}
- - name: megacli.bbu_cycle_count
- description: Cycle Count
- unit: "cycle count"
- chart_type: line
- dimensions:
- - name: adapter {battery id}
diff --git a/collectors/python.d.plugin/memcached/Makefile.inc b/collectors/python.d.plugin/memcached/Makefile.inc
deleted file mode 100644
index e60357161..000000000
--- a/collectors/python.d.plugin/memcached/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += memcached/memcached.chart.py
-dist_pythonconfig_DATA += memcached/memcached.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += memcached/README.md memcached/Makefile.inc
-
diff --git a/collectors/python.d.plugin/memcached/metadata.yaml b/collectors/python.d.plugin/memcached/metadata.yaml
deleted file mode 100644
index 38c9f6853..000000000
--- a/collectors/python.d.plugin/memcached/metadata.yaml
+++ /dev/null
@@ -1,247 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: memcached
- monitored_instance:
- name: Memcached
- link: https://memcached.org/
- categories:
- - data-collection.database-servers
- icon_filename: "memcached.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - memcached
- - memcache
- - cache
- - database
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Monitor Memcached metrics for proficient in-memory key-value store operations. Track cache hits, misses, and memory usage for efficient data caching."
- method_description: "It reads server response to stats command ([stats interface](https://github.com/memcached/memcached/wiki/Commands#stats))."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: |
- If no configuration is given, collector will attempt to connect to memcached instance on `127.0.0.1:11211` address.
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: python.d/memcached.conf
- description: ""
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: Config options
- enabled: true
- list:
- - name: host
- description: the host to connect to.
- default_value: "127.0.0.1"
- required: false
- - name: port
- description: the port to connect to.
- default_value: "11211"
- required: false
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 10
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ""
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: localhost
- description: An example configuration for localhost.
- folding:
- enabled: false
- config: |
- localhost:
- name: 'local'
- host: 'localhost'
- port: 11211
- - name: localipv4
- description: An example configuration for localipv4.
- folding:
- enabled: true
- config: |
- localhost:
- name: 'local'
- host: '127.0.0.1'
- port: 11211
- - name: localipv6
- description: An example configuration for localipv6.
- folding:
- enabled: true
- config: |
- localhost:
- name: 'local'
- host: '::1'
- port: 11211
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: memcached_cache_memory_usage
- link: https://github.com/netdata/netdata/blob/master/health/health.d/memcached.conf
- metric: memcached.cache
- info: cache memory utilization
- - name: memcached_cache_fill_rate
- link: https://github.com/netdata/netdata/blob/master/health/health.d/memcached.conf
- metric: memcached.cache
- info: average rate the cache fills up (positive), or frees up (negative) space over the last hour
- - name: memcached_out_of_cache_space_time
- link: https://github.com/netdata/netdata/blob/master/health/health.d/memcached.conf
- metric: memcached.cache
- info: estimated time the cache will run out of space if the system continues to add data at the same rate as the past hour
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: memcached.cache
- description: Cache Size
- unit: "MiB"
- chart_type: stacked
- dimensions:
- - name: available
- - name: used
- - name: memcached.net
- description: Network
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: in
- - name: out
- - name: memcached.connections
- description: Connections
- unit: "connections/s"
- chart_type: line
- dimensions:
- - name: current
- - name: rejected
- - name: total
- - name: memcached.items
- description: Items
- unit: "items"
- chart_type: line
- dimensions:
- - name: current
- - name: total
- - name: memcached.evicted_reclaimed
- description: Evicted and Reclaimed Items
- unit: "items"
- chart_type: line
- dimensions:
- - name: reclaimed
- - name: evicted
- - name: memcached.get
- description: Get Requests
- unit: "requests"
- chart_type: stacked
- dimensions:
- - name: hints
- - name: misses
- - name: memcached.get_rate
- description: Get Request Rate
- unit: "requests/s"
- chart_type: line
- dimensions:
- - name: rate
- - name: memcached.set_rate
- description: Set Request Rate
- unit: "requests/s"
- chart_type: line
- dimensions:
- - name: rate
- - name: memcached.delete
- description: Delete Requests
- unit: "requests"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: memcached.cas
- description: Check and Set Requests
- unit: "requests"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: bad value
- - name: memcached.increment
- description: Increment Requests
- unit: "requests"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: memcached.decrement
- description: Decrement Requests
- unit: "requests"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: memcached.touch
- description: Touch Requests
- unit: "requests"
- chart_type: stacked
- dimensions:
- - name: hits
- - name: misses
- - name: memcached.touch_rate
- description: Touch Request Rate
- unit: "requests/s"
- chart_type: line
- dimensions:
- - name: rate
diff --git a/collectors/python.d.plugin/monit/Makefile.inc b/collectors/python.d.plugin/monit/Makefile.inc
deleted file mode 100644
index 4a3673fd5..000000000
--- a/collectors/python.d.plugin/monit/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += monit/monit.chart.py
-dist_pythonconfig_DATA += monit/monit.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += monit/README.md monit/Makefile.inc
-
diff --git a/collectors/python.d.plugin/nsd/Makefile.inc b/collectors/python.d.plugin/nsd/Makefile.inc
deleted file mode 100644
index 58e9fd67d..000000000
--- a/collectors/python.d.plugin/nsd/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += nsd/nsd.chart.py
-dist_pythonconfig_DATA += nsd/nsd.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += nsd/README.md nsd/Makefile.inc
-
diff --git a/collectors/python.d.plugin/nvidia_smi/Makefile.inc b/collectors/python.d.plugin/nvidia_smi/Makefile.inc
deleted file mode 100644
index 52fb25a68..000000000
--- a/collectors/python.d.plugin/nvidia_smi/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += nvidia_smi/nvidia_smi.chart.py
-dist_pythonconfig_DATA += nvidia_smi/nvidia_smi.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += nvidia_smi/README.md nvidia_smi/Makefile.inc
-
diff --git a/collectors/python.d.plugin/nvidia_smi/README.md b/collectors/python.d.plugin/nvidia_smi/README.md
deleted file mode 100644
index 7d45289a4..000000000
--- a/collectors/python.d.plugin/nvidia_smi/README.md
+++ /dev/null
@@ -1,157 +0,0 @@
-<!--
-title: "Nvidia GPU monitoring with Netdata"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/nvidia_smi/README.md"
-sidebar_label: "nvidia_smi-python.d.plugin"
-learn_status: "Published"
-learn_topic_type: "References"
-learn_rel_path: "Integrations/Monitor/Devices"
--->
-
-# Nvidia GPU collector
-
-Monitors performance metrics (memory usage, fan speed, pcie bandwidth utilization, temperature, etc.) using `nvidia-smi` cli tool.
-
-## Requirements and Notes
-
-- You must have the `nvidia-smi` tool installed and your NVIDIA GPU(s) must support the tool. Mostly the newer high end models used for AI / ML and Crypto or Pro range, read more about [nvidia_smi](https://developer.nvidia.com/nvidia-system-management-interface).
-- You must enable this plugin, as its disabled by default due to minor performance issues:
- ```bash
- cd /etc/netdata # Replace this path with your Netdata config directory, if different
- sudo ./edit-config python.d.conf
- ```
- Remove the '#' before nvidia_smi so it reads: `nvidia_smi: yes`.
-
-- On some systems when the GPU is idle the `nvidia-smi` tool unloads and there is added latency again when it is next queried. If you are running GPUs under constant workload this isn't likely to be an issue.
-- Currently the `nvidia-smi` tool is being queried via cli. Updating the plugin to use the nvidia c/c++ API directly should resolve this issue. See discussion here: <https://github.com/netdata/netdata/pull/4357>
-- Contributions are welcome.
-- Make sure `netdata` user can execute `/usr/bin/nvidia-smi` or wherever your binary is.
-- If `nvidia-smi` process [is not killed after netdata restart](https://github.com/netdata/netdata/issues/7143) you need to off `loop_mode`.
-- `poll_seconds` is how often in seconds the tool is polled for as an integer.
-
-## Charts
-
-It produces the following charts:
-
-- PCI Express Bandwidth Utilization in `KiB/s`
-- Fan Speed in `percentage`
-- GPU Utilization in `percentage`
-- Memory Bandwidth Utilization in `percentage`
-- Encoder/Decoder Utilization in `percentage`
-- Memory Usage in `MiB`
-- Temperature in `celsius`
-- Clock Frequencies in `MHz`
-- Power Utilization in `Watts`
-- Memory Used by Each Process in `MiB`
-- Memory Used by Each User in `MiB`
-- Number of User on GPU in `num`
-
-## Configuration
-
-Edit the `python.d/nvidia_smi.conf` configuration file using `edit-config` from the Netdata [config
-directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`.
-
-```bash
-cd /etc/netdata # Replace this path with your Netdata config directory, if different
-sudo ./edit-config python.d/nvidia_smi.conf
-```
-
-Sample:
-
-```yaml
-loop_mode : yes
-poll_seconds : 1
-exclude_zero_memory_users : yes
-```
-
-
-### Troubleshooting
-
-To troubleshoot issues with the `nvidia_smi` module, run the `python.d.plugin` with the debug option enabled. The
-output will give you the output of the data collection job or error messages on why the collector isn't working.
-
-First, navigate to your plugins directory, usually they are located under `/usr/libexec/netdata/plugins.d/`. If that's
-not the case on your system, open `netdata.conf` and look for the setting `plugins directory`. Once you're in the
-plugin's directory, switch to the `netdata` user.
-
-```bash
-cd /usr/libexec/netdata/plugins.d/
-sudo su -s /bin/bash netdata
-```
-
-Now you can manually run the `nvidia_smi` module in debug mode:
-
-```bash
-./python.d.plugin nvidia_smi debug trace
-```
-
-## Docker
-
-GPU monitoring in a docker container is possible with [nvidia-container-toolkit](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html) installed on the host system, and `gcompat` added to the `NETDATA_EXTRA_APK_PACKAGES` environment variable.
-
-Sample `docker-compose.yml`
-```yaml
-version: '3'
-services:
- netdata:
- image: netdata/netdata
- container_name: netdata
- hostname: example.com # set to fqdn of host
- ports:
- - 19999:19999
- restart: unless-stopped
- cap_add:
- - SYS_PTRACE
- security_opt:
- - apparmor:unconfined
- environment:
- - NETDATA_EXTRA_APK_PACKAGES=gcompat
- volumes:
- - netdataconfig:/etc/netdata
- - netdatalib:/var/lib/netdata
- - netdatacache:/var/cache/netdata
- - /etc/passwd:/host/etc/passwd:ro
- - /etc/group:/host/etc/group:ro
- - /proc:/host/proc:ro
- - /sys:/host/sys:ro
- - /etc/os-release:/host/etc/os-release:ro
- deploy:
- resources:
- reservations:
- devices:
- - driver: nvidia
- count: all
- capabilities: [gpu]
-
-volumes:
- netdataconfig:
- netdatalib:
- netdatacache:
-```
-
-Sample `docker run`
-```yaml
-docker run -d --name=netdata \
- -p 19999:19999 \
- -e NETDATA_EXTRA_APK_PACKAGES=gcompat \
- -v netdataconfig:/etc/netdata \
- -v netdatalib:/var/lib/netdata \
- -v netdatacache:/var/cache/netdata \
- -v /etc/passwd:/host/etc/passwd:ro \
- -v /etc/group:/host/etc/group:ro \
- -v /proc:/host/proc:ro \
- -v /sys:/host/sys:ro \
- -v /etc/os-release:/host/etc/os-release:ro \
- --restart unless-stopped \
- --cap-add SYS_PTRACE \
- --security-opt apparmor=unconfined \
- --gpus all \
- netdata/netdata
-```
-
-### Docker Troubleshooting
-To troubleshoot `nvidia-smi` in a docker container, first confirm that `nvidia-smi` is working on the host system. If that is working correctly, run `docker exec -it netdata nvidia-smi` to confirm it's working within the docker container. If `nvidia-smi` is fuctioning both inside and outside of the container, confirm that `nvidia-smi: yes` is uncommented in `python.d.conf`.
-```bash
-docker exec -it netdata bash
-cd /etc/netdata
-./edit-config python.d.conf
-```
diff --git a/collectors/python.d.plugin/nvidia_smi/metadata.yaml b/collectors/python.d.plugin/nvidia_smi/metadata.yaml
deleted file mode 100644
index 9bf1e6ca7..000000000
--- a/collectors/python.d.plugin/nvidia_smi/metadata.yaml
+++ /dev/null
@@ -1,166 +0,0 @@
-# This collector will not appear in documentation, as the go version is preferred,
-# https://github.com/netdata/go.d.plugin/blob/master/modules/nvidia_smi/README.md
-#
-# meta:
-# plugin_name: python.d.plugin
-# module_name: nvidia_smi
-# monitored_instance:
-# name: python.d nvidia_smi
-# link: ''
-# categories: []
-# icon_filename: ''
-# related_resources:
-# integrations:
-# list: []
-# info_provided_to_referring_integrations:
-# description: ''
-# keywords: []
-# most_popular: false
-# overview:
-# data_collection:
-# metrics_description: ''
-# method_description: ''
-# supported_platforms:
-# include: []
-# exclude: []
-# multi_instance: true
-# additional_permissions:
-# description: ''
-# default_behavior:
-# auto_detection:
-# description: ''
-# limits:
-# description: ''
-# performance_impact:
-# description: ''
-# setup:
-# prerequisites:
-# list: []
-# configuration:
-# file:
-# name: ''
-# description: ''
-# options:
-# description: ''
-# folding:
-# title: ''
-# enabled: true
-# list: []
-# examples:
-# folding:
-# enabled: true
-# title: ''
-# list: []
-# troubleshooting:
-# problems:
-# list: []
-# alerts: []
-# metrics:
-# folding:
-# title: Metrics
-# enabled: false
-# description: ""
-# availability: []
-# scopes:
-# - name: GPU
-# description: ""
-# labels: []
-# metrics:
-# - name: nvidia_smi.pci_bandwidth
-# description: PCI Express Bandwidth Utilization
-# unit: "KiB/s"
-# chart_type: area
-# dimensions:
-# - name: rx
-# - name: tx
-# - name: nvidia_smi.pci_bandwidth_percent
-# description: PCI Express Bandwidth Percent
-# unit: "percentage"
-# chart_type: area
-# dimensions:
-# - name: rx_percent
-# - name: tx_percent
-# - name: nvidia_smi.fan_speed
-# description: Fan Speed
-# unit: "percentage"
-# chart_type: line
-# dimensions:
-# - name: speed
-# - name: nvidia_smi.gpu_utilization
-# description: GPU Utilization
-# unit: "percentage"
-# chart_type: line
-# dimensions:
-# - name: utilization
-# - name: nvidia_smi.mem_utilization
-# description: Memory Bandwidth Utilization
-# unit: "percentage"
-# chart_type: line
-# dimensions:
-# - name: utilization
-# - name: nvidia_smi.encoder_utilization
-# description: Encoder/Decoder Utilization
-# unit: "percentage"
-# chart_type: line
-# dimensions:
-# - name: encoder
-# - name: decoder
-# - name: nvidia_smi.memory_allocated
-# description: Memory Usage
-# unit: "MiB"
-# chart_type: stacked
-# dimensions:
-# - name: free
-# - name: used
-# - name: nvidia_smi.bar1_memory_usage
-# description: Bar1 Memory Usage
-# unit: "MiB"
-# chart_type: stacked
-# dimensions:
-# - name: free
-# - name: used
-# - name: nvidia_smi.temperature
-# description: Temperature
-# unit: "celsius"
-# chart_type: line
-# dimensions:
-# - name: temp
-# - name: nvidia_smi.clocks
-# description: Clock Frequencies
-# unit: "MHz"
-# chart_type: line
-# dimensions:
-# - name: graphics
-# - name: video
-# - name: sm
-# - name: mem
-# - name: nvidia_smi.power
-# description: Power Utilization
-# unit: "Watts"
-# chart_type: line
-# dimensions:
-# - name: power
-# - name: nvidia_smi.power_state
-# description: Power State
-# unit: "state"
-# chart_type: line
-# dimensions:
-# - name: a dimension per {power_state}
-# - name: nvidia_smi.processes_mem
-# description: Memory Used by Each Process
-# unit: "MiB"
-# chart_type: stacked
-# dimensions:
-# - name: a dimension per process
-# - name: nvidia_smi.user_mem
-# description: Memory Used by Each User
-# unit: "MiB"
-# chart_type: stacked
-# dimensions:
-# - name: a dimension per user
-# - name: nvidia_smi.user_num
-# description: Number of User on GPU
-# unit: "num"
-# chart_type: line
-# dimensions:
-# - name: users
diff --git a/collectors/python.d.plugin/openldap/Makefile.inc b/collectors/python.d.plugin/openldap/Makefile.inc
deleted file mode 100644
index dc947e214..000000000
--- a/collectors/python.d.plugin/openldap/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += openldap/openldap.chart.py
-dist_pythonconfig_DATA += openldap/openldap.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += openldap/README.md openldap/Makefile.inc
-
diff --git a/collectors/python.d.plugin/oracledb/Makefile.inc b/collectors/python.d.plugin/oracledb/Makefile.inc
deleted file mode 100644
index ea3a8240a..000000000
--- a/collectors/python.d.plugin/oracledb/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += oracledb/oracledb.chart.py
-dist_pythonconfig_DATA += oracledb/oracledb.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += oracledb/README.md oracledb/Makefile.inc
-
diff --git a/collectors/python.d.plugin/pandas/Makefile.inc b/collectors/python.d.plugin/pandas/Makefile.inc
deleted file mode 100644
index 9f4f9b34b..000000000
--- a/collectors/python.d.plugin/pandas/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += pandas/pandas.chart.py
-dist_pythonconfig_DATA += pandas/pandas.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += pandas/README.md pandas/Makefile.inc
-
diff --git a/collectors/python.d.plugin/postfix/Makefile.inc b/collectors/python.d.plugin/postfix/Makefile.inc
deleted file mode 100644
index f4091b217..000000000
--- a/collectors/python.d.plugin/postfix/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += postfix/postfix.chart.py
-dist_pythonconfig_DATA += postfix/postfix.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += postfix/README.md postfix/Makefile.inc
-
diff --git a/collectors/python.d.plugin/puppet/Makefile.inc b/collectors/python.d.plugin/puppet/Makefile.inc
deleted file mode 100644
index fe94b9254..000000000
--- a/collectors/python.d.plugin/puppet/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += puppet/puppet.chart.py
-dist_pythonconfig_DATA += puppet/puppet.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += puppet/README.md puppet/Makefile.inc
-
diff --git a/collectors/python.d.plugin/rethinkdbs/Makefile.inc b/collectors/python.d.plugin/rethinkdbs/Makefile.inc
deleted file mode 100644
index dec604464..000000000
--- a/collectors/python.d.plugin/rethinkdbs/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += rethinkdbs/rethinkdbs.chart.py
-dist_pythonconfig_DATA += rethinkdbs/rethinkdbs.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += rethinkdbs/README.md rethinkdbs/Makefile.inc
-
diff --git a/collectors/python.d.plugin/retroshare/Makefile.inc b/collectors/python.d.plugin/retroshare/Makefile.inc
deleted file mode 100644
index 891193e6d..000000000
--- a/collectors/python.d.plugin/retroshare/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += retroshare/retroshare.chart.py
-dist_pythonconfig_DATA += retroshare/retroshare.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += retroshare/README.md retroshare/Makefile.inc
-
diff --git a/collectors/python.d.plugin/retroshare/metadata.yaml b/collectors/python.d.plugin/retroshare/metadata.yaml
deleted file mode 100644
index 0a769616b..000000000
--- a/collectors/python.d.plugin/retroshare/metadata.yaml
+++ /dev/null
@@ -1,144 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: retroshare
- monitored_instance:
- name: RetroShare
- link: "https://retroshare.cc/"
- categories:
- - data-collection.media-streaming-servers
- icon_filename: "retroshare.png"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - retroshare
- - p2p
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This collector monitors RetroShare statistics such as application bandwidth, peers, and DHT metrics."
- method_description: "It connects to the RetroShare web interface to gather metrics."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: "The collector will attempt to connect and detect a RetroShare web interface through http://localhost:9090, even without any configuration."
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: "RetroShare web interface"
- description: |
- RetroShare needs to be configured to enable the RetroShare WEB Interface and allow access from the Netdata host.
- configuration:
- file:
- name: python.d/retroshare.conf
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: name
- description: Job name. This value will overwrite the `job_name` value. JOBS with the same name are mutually exclusive. Only one of them will be allowed running at any time. This allows autodetection to try several alternatives and pick the one that works.
- default_value: ""
- required: false
- - name: url
- description: The URL to the RetroShare Web UI.
- default_value: "http://localhost:9090"
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Local RetroShare Web UI
- description: A basic configuration for a RetroShare server running on localhost.
- config: |
- localhost:
- name: 'local retroshare'
- url: 'http://localhost:9090'
- - name: Remote RetroShare Web UI
- description: A basic configuration for a remote RetroShare server.
- config: |
- remote:
- name: 'remote retroshare'
- url: 'http://1.2.3.4:9090'
-
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: retroshare_dht_working
- link: https://github.com/netdata/netdata/blob/master/health/health.d/retroshare.conf
- metric: retroshare.dht
- info: number of DHT peers
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: retroshare.bandwidth
- description: RetroShare Bandwidth
- unit: "kilobits/s"
- chart_type: area
- dimensions:
- - name: Upload
- - name: Download
- - name: retroshare.peers
- description: RetroShare Peers
- unit: "peers"
- chart_type: line
- dimensions:
- - name: All friends
- - name: Connected friends
- - name: retroshare.dht
- description: Retroshare DHT
- unit: "peers"
- chart_type: line
- dimensions:
- - name: DHT nodes estimated
- - name: RS nodes estimated
diff --git a/collectors/python.d.plugin/riakkv/Makefile.inc b/collectors/python.d.plugin/riakkv/Makefile.inc
deleted file mode 100644
index 87d29f82f..000000000
--- a/collectors/python.d.plugin/riakkv/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += riakkv/riakkv.chart.py
-dist_pythonconfig_DATA += riakkv/riakkv.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += riakkv/README.md riakkv/Makefile.inc
-
diff --git a/collectors/python.d.plugin/riakkv/metadata.yaml b/collectors/python.d.plugin/riakkv/metadata.yaml
deleted file mode 100644
index 441937f89..000000000
--- a/collectors/python.d.plugin/riakkv/metadata.yaml
+++ /dev/null
@@ -1,358 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: riakkv
- monitored_instance:
- name: RiakKV
- link: "https://riak.com/products/riak-kv/index.html"
- categories:
- - data-collection.database-servers
- icon_filename: "riak.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - database
- - nosql
- - big data
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- This collector monitors RiakKV metrics about throughput, latency, resources and more.'
- method_description: "This collector reads the database stats from the `/stats` endpoint."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: "If the /stats endpoint is accessible, RiakKV instances on the local host running on port 8098 will be autodetected."
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: Configure RiakKV to enable /stats endpoint
- description: |
- You can follow the RiakKV configuration reference documentation for how to enable this.
-
- Source : https://docs.riak.com/riak/kv/2.2.3/configuring/reference/#client-interfaces
- configuration:
- file:
- name: "python.d/riakkv.conf"
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- - name: url
- description: The url of the server
- default_value: no
- required: true
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Basic (default)
- folding:
- enabled: false
- description: A basic example configuration per job
- config: |
- local:
- url: 'http://localhost:8098/stats'
- - name: Multi-instance
- description: |
- > **Note**: When you define multiple jobs, their names must be unique.
-
- Collecting metrics from local and remote instances.
- config: |
- local:
- url: 'http://localhost:8098/stats'
-
- remote:
- url: 'http://192.0.2.1:8098/stats'
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: riakkv_1h_kv_get_mean_latency
- link: https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf
- metric: riak.kv.latency.get
- info: average time between reception of client GET request and subsequent response to client over the last hour
- - name: riakkv_kv_get_slow
- link: https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf
- metric: riak.kv.latency.get
- info: average time between reception of client GET request and subsequent response to the client over the last 3 minutes, compared to the average over the last hour
- - name: riakkv_1h_kv_put_mean_latency
- link: https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf
- metric: riak.kv.latency.put
- info: average time between reception of client PUT request and subsequent response to the client over the last hour
- - name: riakkv_kv_put_slow
- link: https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf
- metric: riak.kv.latency.put
- info: average time between reception of client PUT request and subsequent response to the client over the last 3 minutes, compared to the average over the last hour
- - name: riakkv_vm_high_process_count
- link: https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf
- metric: riak.vm
- info: number of processes running in the Erlang VM
- - name: riakkv_list_keys_active
- link: https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf
- metric: riak.core.fsm_active
- info: number of currently running list keys finite state machines
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: riak.kv.throughput
- description: Reads & writes coordinated by this node
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: gets
- - name: puts
- - name: riak.dt.vnode_updates
- description: Update operations coordinated by local vnodes by data type
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: counters
- - name: sets
- - name: maps
- - name: riak.search
- description: Search queries on the node
- unit: "queries/s"
- chart_type: line
- dimensions:
- - name: queries
- - name: riak.search.documents
- description: Documents indexed by search
- unit: "documents/s"
- chart_type: line
- dimensions:
- - name: indexed
- - name: riak.consistent.operations
- description: Consistent node operations
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: gets
- - name: puts
- - name: riak.kv.latency.get
- description: Time between reception of a client GET request and subsequent response to client
- unit: "ms"
- chart_type: line
- dimensions:
- - name: mean
- - name: median
- - name: "95"
- - name: "99"
- - name: "100"
- - name: riak.kv.latency.put
- description: Time between reception of a client PUT request and subsequent response to client
- unit: "ms"
- chart_type: line
- dimensions:
- - name: mean
- - name: median
- - name: "95"
- - name: "99"
- - name: "100"
- - name: riak.dt.latency.counter_merge
- description: Time it takes to perform an Update Counter operation
- unit: "ms"
- chart_type: line
- dimensions:
- - name: mean
- - name: median
- - name: "95"
- - name: "99"
- - name: "100"
- - name: riak.dt.latency.set_merge
- description: Time it takes to perform an Update Set operation
- unit: "ms"
- chart_type: line
- dimensions:
- - name: mean
- - name: median
- - name: "95"
- - name: "99"
- - name: "100"
- - name: riak.dt.latency.map_merge
- description: Time it takes to perform an Update Map operation
- unit: "ms"
- chart_type: line
- dimensions:
- - name: mean
- - name: median
- - name: "95"
- - name: "99"
- - name: "100"
- - name: riak.search.latency.query
- description: Search query latency
- unit: "ms"
- chart_type: line
- dimensions:
- - name: median
- - name: min
- - name: "95"
- - name: "99"
- - name: "999"
- - name: max
- - name: riak.search.latency.index
- description: Time it takes Search to index a new document
- unit: "ms"
- chart_type: line
- dimensions:
- - name: median
- - name: min
- - name: "95"
- - name: "99"
- - name: "999"
- - name: max
- - name: riak.consistent.latency.get
- description: Strongly consistent read latency
- unit: "ms"
- chart_type: line
- dimensions:
- - name: mean
- - name: median
- - name: "95"
- - name: "99"
- - name: "100"
- - name: riak.consistent.latency.put
- description: Strongly consistent write latency
- unit: "ms"
- chart_type: line
- dimensions:
- - name: mean
- - name: median
- - name: "95"
- - name: "99"
- - name: "100"
- - name: riak.vm
- description: Total processes running in the Erlang VM
- unit: "total"
- chart_type: line
- dimensions:
- - name: processes
- - name: riak.vm.memory.processes
- description: Memory allocated & used by Erlang processes
- unit: "MB"
- chart_type: line
- dimensions:
- - name: allocated
- - name: used
- - name: riak.kv.siblings_encountered.get
- description: Number of siblings encountered during GET operations by this node during the past minute
- unit: "siblings"
- chart_type: line
- dimensions:
- - name: mean
- - name: median
- - name: "95"
- - name: "99"
- - name: "100"
- - name: riak.kv.objsize.get
- description: Object size encountered by this node during the past minute
- unit: "KB"
- chart_type: line
- dimensions:
- - name: mean
- - name: median
- - name: "95"
- - name: "99"
- - name: "100"
- - name: riak.search.vnodeq_size
- description: Number of unprocessed messages in the vnode message queues of Search on this node in the past minute
- unit: "messages"
- chart_type: line
- dimensions:
- - name: mean
- - name: median
- - name: "95"
- - name: "99"
- - name: "100"
- - name: riak.search.index
- description: Number of document index errors encountered by Search
- unit: "errors"
- chart_type: line
- dimensions:
- - name: errors
- - name: riak.core.protobuf_connections
- description: Protocol buffer connections by status
- unit: "connections"
- chart_type: line
- dimensions:
- - name: active
- - name: riak.core.repairs
- description: Number of repair operations this node has coordinated
- unit: "repairs"
- chart_type: line
- dimensions:
- - name: read
- - name: riak.core.fsm_active
- description: Active finite state machines by kind
- unit: "fsms"
- chart_type: line
- dimensions:
- - name: get
- - name: put
- - name: secondary index
- - name: list keys
- - name: riak.core.fsm_rejected
- description: Finite state machines being rejected by Sidejobs overload protection
- unit: "fsms"
- chart_type: line
- dimensions:
- - name: get
- - name: put
- - name: riak.search.index
- description: Number of writes to Search failed due to bad data format by reason
- unit: "writes"
- chart_type: line
- dimensions:
- - name: bad_entry
- - name: extract_fail
diff --git a/collectors/python.d.plugin/samba/Makefile.inc b/collectors/python.d.plugin/samba/Makefile.inc
deleted file mode 100644
index 230a8ba43..000000000
--- a/collectors/python.d.plugin/samba/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += samba/samba.chart.py
-dist_pythonconfig_DATA += samba/samba.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += samba/README.md samba/Makefile.inc
-
diff --git a/collectors/python.d.plugin/samba/metadata.yaml b/collectors/python.d.plugin/samba/metadata.yaml
deleted file mode 100644
index ec31e0475..000000000
--- a/collectors/python.d.plugin/samba/metadata.yaml
+++ /dev/null
@@ -1,205 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: samba
- monitored_instance:
- name: Samba
- link: https://www.samba.org/samba/
- categories:
- - data-collection.storage-mount-points-and-filesystems
- icon_filename: "samba.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - samba
- - file sharing
- most_popular: false
- overview:
- data_collection:
- metrics_description: "This collector monitors the performance metrics of Samba file sharing."
- method_description: |
- It is using the `smbstatus` command-line tool.
-
- Executed commands:
-
- - `sudo -n smbstatus -P`
- supported_platforms:
- include: []
- exclude: []
- multi_instance: false
- additional_permissions:
- description: |
- `smbstatus` is used, which can only be executed by `root`. It uses `sudo` and assumes that it is configured such that the `netdata` user can execute `smbstatus` as root without a password.
- default_behavior:
- auto_detection:
- description: "After all the permissions are satisfied, the `smbstatus -P` binary is executed."
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: Enable the samba collector
- description: |
- The `samba` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`, to edit the `python.d.conf` file.
-
- ```bash
- cd /etc/netdata # Replace this path with your Netdata config directory, if different
- sudo ./edit-config python.d.conf
- ```
- Change the value of the `samba` setting to `yes`. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
- - title: Permissions and programs
- description: |
- To run the collector you need:
-
- - `smbstatus` program
- - `sudo` program
- - `smbd` must be compiled with profiling enabled
- - `smbd` must be started either with the `-P 1` option or inside `smb.conf` using `smbd profiling level`
-
- The module uses `smbstatus`, which can only be executed by `root`. It uses `sudo` and assumes that it is configured such that the `netdata` user can execute `smbstatus` as root without a password.
-
- - add to your `/etc/sudoers` file:
-
- `which smbstatus` shows the full path to the binary.
-
- ```bash
- netdata ALL=(root) NOPASSWD: /path/to/smbstatus
- ```
-
- - Reset Netdata's systemd unit [CapabilityBoundingSet](https://www.freedesktop.org/software/systemd/man/systemd.exec.html#Capabilities) (Linux distributions with systemd)
-
- The default CapabilityBoundingSet doesn't allow using `sudo`, and is quite strict in general. Resetting is not optimal, but a next-best solution given the inability to execute `smbstatus` using `sudo`.
-
-
- As the `root` user, do the following:
-
- ```cmd
- mkdir /etc/systemd/system/netdata.service.d
- echo -e '[Service]\nCapabilityBoundingSet=~' | tee /etc/systemd/system/netdata.service.d/unset-capability-bounding-set.conf
- systemctl daemon-reload
- systemctl restart netdata.service
- ```
- configuration:
- file:
- name: python.d/samba.conf
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Basic
- description: A basic example configuration.
- config: |
- my_job_name:
- name: my_name
- update_every: 1
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: syscall.rw
- description: R/Ws
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: sendfile
- - name: recvfile
- - name: smb2.rw
- description: R/Ws
- unit: "KiB/s"
- chart_type: area
- dimensions:
- - name: readout
- - name: writein
- - name: readin
- - name: writeout
- - name: smb2.create_close
- description: Create/Close
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: create
- - name: close
- - name: smb2.get_set_info
- description: Info
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: getinfo
- - name: setinfo
- - name: smb2.find
- description: Find
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: find
- - name: smb2.notify
- description: Notify
- unit: "operations/s"
- chart_type: line
- dimensions:
- - name: notify
- - name: smb2.sm_counters
- description: Lesser Ops
- unit: "count"
- chart_type: stacked
- dimensions:
- - name: tcon
- - name: negprot
- - name: tdis
- - name: cancel
- - name: logoff
- - name: flush
- - name: lock
- - name: keepalive
- - name: break
- - name: sessetup
diff --git a/collectors/python.d.plugin/sensors/Makefile.inc b/collectors/python.d.plugin/sensors/Makefile.inc
deleted file mode 100644
index 5fb26e1c8..000000000
--- a/collectors/python.d.plugin/sensors/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += sensors/sensors.chart.py
-dist_pythonconfig_DATA += sensors/sensors.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += sensors/README.md sensors/Makefile.inc
-
diff --git a/collectors/python.d.plugin/sensors/metadata.yaml b/collectors/python.d.plugin/sensors/metadata.yaml
deleted file mode 100644
index d7cb2206f..000000000
--- a/collectors/python.d.plugin/sensors/metadata.yaml
+++ /dev/null
@@ -1,184 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: sensors
- monitored_instance:
- name: Linux Sensors (lm-sensors)
- link: https://hwmon.wiki.kernel.org/lm_sensors
- categories:
- - data-collection.hardware-devices-and-sensors
- icon_filename: "microchip.svg"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - sensors
- - temperature
- - voltage
- - current
- - power
- - fan
- - energy
- - humidity
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- Examine Linux Sensors metrics with Netdata for insights into hardware health and performance.
-
- Enhance your system's reliability with real-time hardware health insights.
- method_description: >
- Reads system sensors information (temperature, voltage, electric current, power, etc.) via [lm-sensors](https://hwmon.wiki.kernel.org/lm_sensors).
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: >
- The following type of sensors are auto-detected:
-
- - temperature
- - fan
- - voltage
- - current
- - power
- - energy
- - humidity
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: python.d/sensors.conf
- description: ""
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: Config options
- enabled: true
- list:
- - name: types
- description: The types of sensors to collect.
- default_value: "temperature, fan, voltage, current, power, energy, humidity"
- required: true
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 1
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: Config
- list:
- - name: Default
- folding:
- enabled: false
- description: Default configuration.
- config: |
- types:
- - temperature
- - fan
- - voltage
- - current
- - power
- - energy
- - humidity
- troubleshooting:
- problems:
- list:
- - name: lm-sensors doesn't work on your device
- description: |
- When `lm-sensors` doesn't work on your device (e.g. for RPi temperatures),
- use [the legacy bash collector](https://github.com/netdata/netdata/blob/master/collectors/charts.d.plugin/sensors/README.md)
- - name: ACPI ring buffer errors are printed
- description: |
- There have been reports from users that on certain servers, ACPI ring buffer errors are printed by the kernel (`dmesg`)
- when ACPI sensors are being accessed. We are tracking such cases in issue [#827](https://github.com/netdata/netdata/issues/827).
- Please join this discussion for help.
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: chip
- description: >
- Metrics related to chips. Each chip provides a set of the following metrics, each having the chip name in the metric name as reported by `sensors -u`.
- labels: []
- metrics:
- - name: sensors.temperature
- description: Temperature
- unit: "Celsius"
- chart_type: line
- dimensions:
- - name: a dimension per sensor
- - name: sensors.voltage
- description: Voltage
- unit: "Volts"
- chart_type: line
- dimensions:
- - name: a dimension per sensor
- - name: sensors.current
- description: Current
- unit: "Ampere"
- chart_type: line
- dimensions:
- - name: a dimension per sensor
- - name: sensors.power
- description: Power
- unit: "Watt"
- chart_type: line
- dimensions:
- - name: a dimension per sensor
- - name: sensors.fan
- description: Fans speed
- unit: "Rotations/min"
- chart_type: line
- dimensions:
- - name: a dimension per sensor
- - name: sensors.energy
- description: Energy
- unit: "Joule"
- chart_type: line
- dimensions:
- - name: a dimension per sensor
- - name: sensors.humidity
- description: Humidity
- unit: "Percent"
- chart_type: line
- dimensions:
- - name: a dimension per sensor
diff --git a/collectors/python.d.plugin/smartd_log/Makefile.inc b/collectors/python.d.plugin/smartd_log/Makefile.inc
deleted file mode 100644
index dc1d0f3fb..000000000
--- a/collectors/python.d.plugin/smartd_log/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += smartd_log/smartd_log.chart.py
-dist_pythonconfig_DATA += smartd_log/smartd_log.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += smartd_log/README.md smartd_log/Makefile.inc
-
diff --git a/collectors/python.d.plugin/spigotmc/Makefile.inc b/collectors/python.d.plugin/spigotmc/Makefile.inc
deleted file mode 100644
index f9fa8b6b0..000000000
--- a/collectors/python.d.plugin/spigotmc/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += spigotmc/spigotmc.chart.py
-dist_pythonconfig_DATA += spigotmc/spigotmc.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += spigotmc/README.md spigotmc/Makefile.inc
-
diff --git a/collectors/python.d.plugin/squid/Makefile.inc b/collectors/python.d.plugin/squid/Makefile.inc
deleted file mode 100644
index 76ecff81e..000000000
--- a/collectors/python.d.plugin/squid/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += squid/squid.chart.py
-dist_pythonconfig_DATA += squid/squid.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += squid/README.md squid/Makefile.inc
-
diff --git a/collectors/python.d.plugin/tomcat/Makefile.inc b/collectors/python.d.plugin/tomcat/Makefile.inc
deleted file mode 100644
index 940a7835e..000000000
--- a/collectors/python.d.plugin/tomcat/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += tomcat/tomcat.chart.py
-dist_pythonconfig_DATA += tomcat/tomcat.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += tomcat/README.md tomcat/Makefile.inc
-
diff --git a/collectors/python.d.plugin/tor/Makefile.inc b/collectors/python.d.plugin/tor/Makefile.inc
deleted file mode 100644
index 5a45f9b79..000000000
--- a/collectors/python.d.plugin/tor/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += tor/tor.chart.py
-dist_pythonconfig_DATA += tor/tor.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += tor/README.md tor/Makefile.inc
-
diff --git a/collectors/python.d.plugin/traefik/Makefile.inc b/collectors/python.d.plugin/traefik/Makefile.inc
deleted file mode 100644
index 926d56dda..000000000
--- a/collectors/python.d.plugin/traefik/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += traefik/traefik.chart.py
-dist_pythonconfig_DATA += traefik/traefik.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += traefik/README.md traefik/Makefile.inc
-
diff --git a/collectors/python.d.plugin/traefik/README.md b/collectors/python.d.plugin/traefik/README.md
deleted file mode 100644
index 40ed24f04..000000000
--- a/collectors/python.d.plugin/traefik/README.md
+++ /dev/null
@@ -1,98 +0,0 @@
-<!--
-title: "Traefik monitoring with Netdata"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/traefik/README.md"
-sidebar_label: "traefik-python.d.plugin"
-learn_status: "Published"
-learn_topic_type: "References"
-learn_rel_path: "Integrations/Monitor/Webapps"
--->
-
-# Traefik collector
-
-Uses the `health` API to provide statistics.
-
-It produces:
-
-1. **Responses** by statuses
-
- - success (1xx, 2xx, 304)
- - error (5xx)
- - redirect (3xx except 304)
- - bad (4xx)
- - other (all other responses)
-
-2. **Responses** by codes
-
- - 2xx (successful)
- - 5xx (internal server errors)
- - 3xx (redirect)
- - 4xx (bad)
- - 1xx (informational)
- - other (non-standart responses)
-
-3. **Detailed Response Codes** requests/s (number of responses for each response code family individually)
-
-4. **Requests**/s
-
- - request statistics
-
-5. **Total response time**
-
- - sum of all response time
-
-6. **Average response time**
-
-7. **Average response time per iteration**
-
-8. **Uptime**
-
- - Traefik server uptime
-
-## Configuration
-
-Edit the `python.d/traefik.conf` configuration file using `edit-config` from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically
-at `/etc/netdata`.
-
-```bash
-cd /etc/netdata # Replace this path with your Netdata config directory, if different
-sudo ./edit-config python.d/traefik.conf
-```
-
-Needs only `url` to server's `health`
-
-Here is an example for local server:
-
-```yaml
-update_every: 1
-priority: 60000
-
-local:
- url: 'http://localhost:8080/health'
-```
-
-Without configuration, module attempts to connect to `http://localhost:8080/health`.
-
-
-
-
-### Troubleshooting
-
-To troubleshoot issues with the `traefik` module, run the `python.d.plugin` with the debug option enabled. The
-output will give you the output of the data collection job or error messages on why the collector isn't working.
-
-First, navigate to your plugins directory, usually they are located under `/usr/libexec/netdata/plugins.d/`. If that's
-not the case on your system, open `netdata.conf` and look for the setting `plugins directory`. Once you're in the
-plugin's directory, switch to the `netdata` user.
-
-```bash
-cd /usr/libexec/netdata/plugins.d/
-sudo su -s /bin/bash netdata
-```
-
-Now you can manually run the `traefik` module in debug mode:
-
-```bash
-./python.d.plugin traefik debug trace
-```
-
diff --git a/collectors/python.d.plugin/traefik/metadata.yaml b/collectors/python.d.plugin/traefik/metadata.yaml
deleted file mode 100644
index dcfb098a0..000000000
--- a/collectors/python.d.plugin/traefik/metadata.yaml
+++ /dev/null
@@ -1,125 +0,0 @@
-# This collector will not appear in documentation, as the go version is preferred,
-# https://github.com/netdata/go.d.plugin/blob/master/modules/traefik/README.md
-#
-# meta:
-# plugin_name: python.d.plugin
-# module_name: traefik
-# monitored_instance:
-# name: python.d traefik
-# link: ''
-# categories: []
-# icon_filename: ''
-# related_resources:
-# integrations:
-# list: []
-# info_provided_to_referring_integrations:
-# description: ''
-# keywords: []
-# most_popular: false
-# overview:
-# data_collection:
-# metrics_description: ''
-# method_description: ''
-# supported_platforms:
-# include: []
-# exclude: []
-# multi_instance: true
-# additional_permissions:
-# description: ''
-# default_behavior:
-# auto_detection:
-# description: ''
-# limits:
-# description: ''
-# performance_impact:
-# description: ''
-# setup:
-# prerequisites:
-# list: []
-# configuration:
-# file:
-# name: ''
-# description: ''
-# options:
-# description: ''
-# folding:
-# title: ''
-# enabled: true
-# list: []
-# examples:
-# folding:
-# enabled: true
-# title: ''
-# list: []
-# troubleshooting:
-# problems:
-# list: []
-# alerts: []
-# metrics:
-# folding:
-# title: Metrics
-# enabled: false
-# description: ""
-# availability: []
-# scopes:
-# - name: global
-# description: ""
-# labels: []
-# metrics:
-# - name: traefik.response_statuses
-# description: Response statuses
-# unit: "requests/s"
-# chart_type: stacked
-# dimensions:
-# - name: success
-# - name: error
-# - name: redirect
-# - name: bad
-# - name: other
-# - name: traefik.response_codes
-# description: Responses by codes
-# unit: "requests/s"
-# chart_type: stacked
-# dimensions:
-# - name: 2xx
-# - name: 5xx
-# - name: 3xx
-# - name: 4xx
-# - name: 1xx
-# - name: other
-# - name: traefik.detailed_response_codes
-# description: Detailed response codes
-# unit: "requests/s"
-# chart_type: stacked
-# dimensions:
-# - name: a dimension for each response code family
-# - name: traefik.requests
-# description: Requests
-# unit: "requests/s"
-# chart_type: line
-# dimensions:
-# - name: requests
-# - name: traefik.total_response_time
-# description: Total response time
-# unit: "seconds"
-# chart_type: line
-# dimensions:
-# - name: response
-# - name: traefik.average_response_time
-# description: Average response time
-# unit: "milliseconds"
-# chart_type: line
-# dimensions:
-# - name: response
-# - name: traefik.average_response_time_per_iteration
-# description: Average response time per iteration
-# unit: "milliseconds"
-# chart_type: line
-# dimensions:
-# - name: response
-# - name: traefik.uptime
-# description: Uptime
-# unit: "seconds"
-# chart_type: line
-# dimensions:
-# - name: uptime
diff --git a/collectors/python.d.plugin/uwsgi/Makefile.inc b/collectors/python.d.plugin/uwsgi/Makefile.inc
deleted file mode 100644
index 75d96de0e..000000000
--- a/collectors/python.d.plugin/uwsgi/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += uwsgi/uwsgi.chart.py
-dist_pythonconfig_DATA += uwsgi/uwsgi.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += uwsgi/README.md uwsgi/Makefile.inc
-
diff --git a/collectors/python.d.plugin/varnish/Makefile.inc b/collectors/python.d.plugin/varnish/Makefile.inc
deleted file mode 100644
index 2469b0592..000000000
--- a/collectors/python.d.plugin/varnish/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += varnish/varnish.chart.py
-dist_pythonconfig_DATA += varnish/varnish.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += varnish/README.md varnish/Makefile.inc
-
diff --git a/collectors/python.d.plugin/w1sensor/Makefile.inc b/collectors/python.d.plugin/w1sensor/Makefile.inc
deleted file mode 100644
index bddf146f5..000000000
--- a/collectors/python.d.plugin/w1sensor/Makefile.inc
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += w1sensor/w1sensor.chart.py
-dist_pythonconfig_DATA += w1sensor/w1sensor.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += w1sensor/README.md w1sensor/Makefile.inc
-
diff --git a/collectors/python.d.plugin/zscores/Makefile.inc b/collectors/python.d.plugin/zscores/Makefile.inc
deleted file mode 100644
index d8b182415..000000000
--- a/collectors/python.d.plugin/zscores/Makefile.inc
+++ /dev/null
@@ -1,12 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-# THIS IS NOT A COMPLETE Makefile
-# IT IS INCLUDED BY ITS PARENT'S Makefile.am
-# IT IS REQUIRED TO REFERENCE ALL FILES RELATIVE TO THE PARENT
-
-# install these files
-dist_python_DATA += zscores/zscores.chart.py
-dist_pythonconfig_DATA += zscores/zscores.conf
-
-# do not install these files, but include them in the distribution
-dist_noinst_DATA += zscores/README.md zscores/Makefile.inc
diff --git a/collectors/python.d.plugin/zscores/metadata.yaml b/collectors/python.d.plugin/zscores/metadata.yaml
deleted file mode 100644
index 388e9b460..000000000
--- a/collectors/python.d.plugin/zscores/metadata.yaml
+++ /dev/null
@@ -1,187 +0,0 @@
-plugin_name: python.d.plugin
-modules:
- - meta:
- plugin_name: python.d.plugin
- module_name: zscores
- monitored_instance:
- name: python.d zscores
- link: https://en.wikipedia.org/wiki/Standard_score
- categories:
- - data-collection.other
- icon_filename: ""
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords:
- - zscore
- - z-score
- - standard score
- - standard deviation
- - anomaly detection
- - statistical anomaly detection
- most_popular: false
- overview:
- data_collection:
- metrics_description: |
- By using smoothed, rolling [Z-Scores](https://en.wikipedia.org/wiki/Standard_score) for selected metrics or charts you can narrow down your focus and shorten root cause analysis.
- method_description: |
- This collector uses the [Netdata rest api](https://github.com/netdata/netdata/blob/master/web/api/README.md) to get the `mean` and `stddev`
- for each dimension on specified charts over a time range (defined by `train_secs` and `offset_secs`).
-
- For each dimension it will calculate a Z-Score as `z = (x - mean) / stddev` (clipped at `z_clip`). Scores are then smoothed over
- time (`z_smooth_n`) and, if `mode: 'per_chart'`, aggregated across dimensions to a smoothed, rolling chart level Z-Score at each time step.
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list:
- - title: Python Requirements
- description: |
- This collector will only work with Python 3 and requires the below packages be installed.
-
- ```bash
- # become netdata user
- sudo su -s /bin/bash netdata
- # install required packages
- pip3 install numpy pandas requests netdata-pandas==0.0.38
- ```
- configuration:
- file:
- name: python.d/zscores.conf
- description: ""
- options:
- description: |
- There are 2 sections:
-
- * Global variables
- * One or more JOBS that can define multiple different instances to monitor.
-
- The following options can be defined globally: priority, penalty, autodetection_retry, update_every, but can also be defined per JOB to override the global values.
-
- Additionally, the following collapsed table contains all the options that can be configured inside a JOB definition.
-
- Every configuration JOB starts with a `job_name` value which will appear in the dashboard, unless a `name` parameter is specified.
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: charts_regex
- description: what charts to pull data for - A regex like `system\..*|` or `system\..*|apps.cpu|apps.mem` etc.
- default_value: "system\\..*"
- required: true
- - name: train_secs
- description: length of time (in seconds) to base calculations off for mean and stddev.
- default_value: 14400
- required: true
- - name: offset_secs
- description: offset (in seconds) preceding latest data to ignore when calculating mean and stddev.
- default_value: 300
- required: true
- - name: train_every_n
- description: recalculate the mean and stddev every n steps of the collector.
- default_value: 900
- required: true
- - name: z_smooth_n
- description: smooth the z score (to reduce sensitivity to spikes) by averaging it over last n values.
- default_value: 15
- required: true
- - name: z_clip
- description: cap absolute value of zscore (before smoothing) for better stability.
- default_value: 10
- required: true
- - name: z_abs
- description: "set z_abs: 'true' to make all zscores be absolute values only."
- default_value: "true"
- required: true
- - name: burn_in
- description: burn in period in which to initially calculate mean and stddev on every step.
- default_value: 2
- required: true
- - name: mode
- description: mode can be to get a zscore 'per_dim' or 'per_chart'.
- default_value: per_chart
- required: true
- - name: per_chart_agg
- description: per_chart_agg is how you aggregate from dimension to chart when mode='per_chart'.
- default_value: mean
- required: true
- - name: update_every
- description: Sets the default data collection frequency.
- default_value: 5
- required: false
- - name: priority
- description: Controls the order of charts at the netdata dashboard.
- default_value: 60000
- required: false
- - name: autodetection_retry
- description: Sets the job re-check interval in seconds.
- default_value: 0
- required: false
- - name: penalty
- description: Indicates whether to apply penalty to update_every in case of failures.
- default_value: yes
- required: false
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Default
- description: Default configuration.
- folding:
- enabled: false
- config: |
- local:
- name: 'local'
- host: '127.0.0.1:19999'
- charts_regex: 'system\..*'
- charts_to_exclude: 'system.uptime'
- train_secs: 14400
- offset_secs: 300
- train_every_n: 900
- z_smooth_n: 15
- z_clip: 10
- z_abs: 'true'
- burn_in: 2
- mode: 'per_chart'
- per_chart_agg: 'mean'
- troubleshooting:
- problems:
- list: []
- alerts: []
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: zscores.z
- description: Z Score
- unit: "z"
- chart_type: line
- dimensions:
- - name: a dimension per chart or dimension
- - name: zscores.3stddev
- description: Z Score >3
- unit: "count"
- chart_type: stacked
- dimensions:
- - name: a dimension per chart or dimension
diff --git a/collectors/slabinfo.plugin/Makefile.am b/collectors/slabinfo.plugin/Makefile.am
deleted file mode 100644
index 07796ea7b..000000000
--- a/collectors/slabinfo.plugin/Makefile.am
+++ /dev/null
@@ -1,14 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-CLEANFILES = \
- slabinfo.plugin \
- $(NULL)
-
-include $(top_srcdir)/build/subst.inc
-SUFFIXES = .in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/statsd.plugin/Makefile.am b/collectors/statsd.plugin/Makefile.am
deleted file mode 100644
index c8144c137..000000000
--- a/collectors/statsd.plugin/Makefile.am
+++ /dev/null
@@ -1,23 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
-
-statsdconfigdir=$(libconfigdir)/statsd.d
-dist_statsdconfig_DATA = \
- example.conf \
- k6.conf \
- asterisk.conf \
- $(NULL)
-
-userstatsdconfigdir=$(configdir)/statsd.d
-dist_userstatsdconfig_DATA = \
- $(NULL)
-
-# Explicitly install directories to avoid permission issues due to umask
-install-exec-local:
- $(INSTALL) -d $(DESTDIR)$(userstatsdconfigdir)
diff --git a/collectors/statsd.plugin/README.md b/collectors/statsd.plugin/README.md
deleted file mode 100644
index e3c8f9f81..000000000
--- a/collectors/statsd.plugin/README.md
+++ /dev/null
@@ -1,1052 +0,0 @@
-<!--
-title: "StatsD"
-description: "The Netdata Agent is a fully-featured StatsD server that collects metrics from any custom application and visualizes them in real-time."
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/statsd.plugin/README.md"
-sidebar_label: "StatsD"
-learn_status: "Published"
-learn_rel_path: "Integrations/Monitor/Anything"
--->
-
-# StatsD
-
-[StatsD](https://github.com/statsd/statsd) is a system to collect data from any application. Applications send metrics to it,
-usually via non-blocking UDP communication, and StatsD servers collect these metrics, perform a few simple calculations on
-them and push them to backend time-series databases.
-
-If you want to learn more about the StatsD protocol, we have written a
-[blog post](https://blog.netdata.cloud/introduction-to-statsd/) about it!
-
-
-Netdata is a fully featured statsd server. It can collect statsd formatted metrics, visualize
-them on its dashboards and store them in it's database for long-term retention.
-
-Netdata statsd is inside Netdata (an internal plugin, running inside the Netdata daemon), it is
-configured via `netdata.conf` and by-default listens on standard statsd port 8125. Netdata supports
-both TCP and UDP packets at the same time.
-
-Since statsd is embedded in Netdata, it means you now have a statsd server embedded on all your servers.
-
-Netdata statsd is fast. It can collect several millions of metrics per second on modern hardware, using
-just 1 CPU core. The implementation uses two threads: one thread collects metrics, another thread updates
-the charts from the collected data.
-
-## Available StatsD synthetic application charts
-
-Netdata ships with a few synthetic chart definitions to automatically present application metrics into a
-more uniform way. These synthetic charts are configuration files (you can create your own) that re-arrange
-statsd metrics into a more meaningful way.
-
-On synthetic charts, we can have alerts as with any metric and chart.
-
-- [K6 load testing tool](https://k6.io)
- - **Description:** k6 is a developer-centric, free and open-source load testing tool built for making performance testing a productive and enjoyable experience.
- - [Documentation](https://github.com/netdata/netdata/blob/master/collectors/statsd.plugin/k6.md)
- - [Configuration](https://github.com/netdata/netdata/blob/master/collectors/statsd.plugin/k6.conf)
-- [Asterisk](https://www.asterisk.org/)
- - **Description:** Asterisk is an Open Source PBX and telephony toolkit.
- - [Documentation](https://github.com/netdata/netdata/blob/master/collectors/statsd.plugin/asterisk.md)
- - [Configuration](https://github.com/netdata/netdata/blob/master/collectors/statsd.plugin/asterisk.conf)
-
-## Metrics supported by Netdata
-
-Netdata fully supports the StatsD protocol and also extends it to support more advanced Netdata specific use cases.
-All StatsD client libraries can be used with Netdata too.
-
-- **Gauges**
-
- The application sends `name:value|g`, where `value` is any **decimal/fractional** number, StatsD reports the
- latest value collected and the number of times it was updated (events).
-
- The application may increment or decrement a previous value, by setting the first character of the value to
- `+` or `-` (so, the only way to set a gauge to an absolute negative value, is to first set it to zero).
-
- [Sampling rate](#sampling-rates) is supported.
- [Tags](#tags) are supported for changing chart units, family and dimension name.
-
- When a gauge is not collected and the setting is not to show gaps on the charts (the default), the last value will be shown, until a data collection event changes it.
-
-- **Counters** and **Meters**
-
- The application sends `name:value|c`, `name:value|C` or `name:value|m`, where `value` is a positive or negative **integer** number of events occurred, StatsD reports the **rate** and the number of times it was updated (events).
-
- `:value` can be omitted and StatsD will assume it is `1`. `|c`, `|C` and `|m` can be omitted and StatsD will assume it is `|m`. So, the application may send just `name` and StatsD will parse it as `name:1|m`.
-
- - Counters use `|c` (etsy/StatsD compatible) or `|C` (brubeck compatible)
- - Meters use `|m`
-
- [Sampling rate](#sampling-rates) is supported.
- [Tags](#tags) are supported for changing chart units, family and dimension name.
-
- When a counter or meter is not collected, StatsD **defaults** to showing a zero value, until a data collection event changes the value.
-
-- **Timers** and **Histograms**
-
- The application sends `name:value|ms` or `name:value|h`, where `value` is any **decimal/fractional** number, StatsD reports **min**, **max**, **average**, **95th percentile**, **median** and **standard deviation** and the total number of times it was updated (events). Internally it also calculates the **sum**, which is available for synthetic charts.
-
- - Timers use `|ms`
- - Histograms use `|h`
-
- The only difference between the two, is the `units` of the charts, as timers report *milliseconds*.
-
- [Sampling rate](#sampling-rates) is supported.
- [Tags](#tags) are supported for changing chart units and family.
-
- When a counter or meter is not collected, StatsD **defaults** to showing a zero value, until a data collection event changes the value.
-
-- **Sets**
-
- The application sends `name:value|s`, where `value` is anything (**number or text**, leading and trailing spaces are removed), StatsD reports the number of unique values sent and the number of times it was updated (events).
-
- Sampling rate is **not** supported for Sets. `value` is always considered text (so `01` and `1` are considered different).
-
- [Tags](#tags) are supported for changing chart units and family.
-
- When a set is not collected, Netdata **defaults** to showing a zero value, until a data collection event changes the value.
-
-- **Dictionaries**
-
- The application sends `name:value|d`, where `value` is anything (**number or text**, leading and trailing spaces are removed), StatsD reports the number of events sent for each `value` and the total times `name` was updated (events).
-
- Sampling rate is **not** supported for Dictionaries. `value` is always considered text (so `01` and `1` are considered different).
-
- [Tags](#tags) are supported for changing chart units and family.
-
- When a set is not collected, Netdata **defaults** to showing a zero value, until a data collection event changes the value.
-
-#### Sampling Rates
-
-The application may append `|@sampling_rate`, where `sampling_rate` is a number from `0.0` to `1.0` in order for StatD to extrapolate the value and predict the total for the entire period. If the application reports to StatsD a value for 1/10th of the time, it can append `|@0.1` to the metrics it sends to statsd.
-
-#### Tags
-
-The application may append `|#tag1:value1,tag2:value2,tag3:value3` etc, where `tagX` and `valueX` are strings. `:valueX` can be omitted.
-
-Currently, Netdata uses only 2 tags:
-
- * `units=string` which sets the units of the chart that is automatically generated
- * `family=string` which sets the family of the chart that is automatically generated (the family is the submenu of the dashboard)
- * `name=string` which sets the name of the dimension of the chart that is automatically generated (only for counters, meters, gauges)
-
-Other tags are parsed, but currently are ignored.
-
-Charts are not updated to change units or dimension names once they are created. So, either send the tags on every event, or use the special `zinit` value to initiaze the charts at the beginning. `zinit` is a special value that can be used on any chart, to have netdata initialize the charts, without actually setting any values to them. So, instead of sending `my.metric:VALUE|c|#units=bytes,name=size` every time, the application can send at the beginning `my.metric:zinit|c|#units=bytes,name=size` and then `my.metric:VALUE|c`.
-
-#### Overlapping metrics
-
-Netdata's StatsD server maintains different indexes for each of the metric types supported. This means the same metric `name` may exist under different types concurrently.
-
-#### How to name your metrics
-
-A good practice is to name your metrics like `application.operation.metric`, where:
-
-- `application` is the application name - Netdata will automatically create a dashboard section based on the first keyword of the metrics, so you can have all your applications in different sections.
-- `operation` is the operation your application is executing, like `dbquery`, `request`, `response`, etc.
-- `metric` is anything you want to name your metric as. Netdata will automatically append the metric type (meter, counter, gauge, set, dictionary, timer, histogram) to the generated chart.
-
-Using [Tags](#tags) you can also change the submenus of the dashboard, the units of the charts and for meters, counters and gauges, the name of dimension. So, you can have a usable default view without using [Synthetic StatsD charts](#synthetic-statsd-charts)
-
-#### Multiple metrics per packet
-
-Netdata accepts multiple metrics per packet if each is terminated with a newline (`\n`) at the end.
-
-#### TCP packets
-
-Netdata listens for both TCP and UDP packets. For TCP, is it important to always append `\n` on each metric, as Netdata will use the newline character to detect if a metric is split into multiple TCP packets.
-
-
-#### UDP packets
-
-When sending multiple metrics over a single UDP message, it is important not to exceed the network MTU, which is usually 1500 bytes.
-
-Netdata will accept UDP packets up to 9000 bytes, but the underlying network will not exceed MTU.
-
-> You can read more about the network maximum transmission unit(MTU) in this cloudflare [article](https://www.cloudflare.com/en-gb/learning/network-layer/what-is-mtu/).
-
-## Configuration
-
-You can find the configuration at `/etc/netdata/netdata.conf`:
-
-```
-[statsd]
- # enabled = yes
- # decimal detail = 1000
- # update every (flushInterval) = 1
- # udp messages to process at once = 10
- # create private charts for metrics matching = *
- # max private charts hard limit = 1000
- # cleanup obsolete charts after secs = 0
- # private charts memory mode = save
- # private charts history = 3996
- # histograms and timers percentile (percentThreshold) = 95.00000
- # add dimension for number of events received = no
- # gaps on gauges (deleteGauges) = no
- # gaps on counters (deleteCounters) = no
- # gaps on meters (deleteMeters) = no
- # gaps on sets (deleteSets) = no
- # gaps on histograms (deleteHistograms) = no
- # gaps on timers (deleteTimers) = no
- # listen backlog = 4096
- # default port = 8125
- # bind to = udp:localhost:8125 tcp:localhost:8125
-```
-
-### StatsD main config options
-
-- `enabled = yes|no`
-
- controls if StatsD will be enabled for this Netdata. The default is enabled.
-
-- `default port = 8125`
-
- controls the default port StatsD will use if no port is defined in the following setting.
-
-- `bind to = udp:localhost tcp:localhost`
-
- is a space separated list of IPs and ports to listen to. The format is `PROTOCOL:IP:PORT` - if `PORT` is omitted, the `default port` will be used. If `IP` is IPv6, it needs to be enclosed in `[]`. `IP` can also be `*` (to listen on all IPs) or even a hostname.
-
-- `update every (flushInterval) = 1` seconds, controls the frequency StatsD will push the collected metrics to Netdata charts.
-
-- `decimal detail = 1000` controls the number of fractional digits in gauges and histograms. Netdata collects metrics using signed 64-bit integers and their fractional detail is controlled using multipliers and divisors. This setting is used to multiply all collected values to convert them to integers and is also set as the divisors, so that the final data will be a floating point number with this fractional detail (1000 = X.0 - X.999, 10000 = X.0 - X.9999, etc).
-
-The rest of the settings are discussed below.
-
-## StatsD charts
-
-Netdata can visualize StatsD collected metrics in 2 ways:
-
-1. Each metric gets its own **private chart**. This is the default and does not require any configuration. You can adjust the default parameters.
-
-2. **Synthetic charts** can be created, combining multiple metrics, independently of their metric types. For this type of charts, special configuration is required, to define the chart title, type, units, its dimensions, etc.
-
-### Private metric charts
-
-Private charts are controlled with `create private charts for metrics matching = *`. This setting accepts a space-separated list of [simple patterns](https://github.com/netdata/netdata/blob/master/libnetdata/simple_pattern/README.md). Netdata will create private charts for all metrics **by default**.
-
-For example, to render charts for all `myapp.*` metrics, except `myapp.*.badmetric`, use:
-
-```
-create private charts for metrics matching = !myapp.*.badmetric myapp.*
-```
-
-You can specify Netdata StatsD to have a different `memory mode` than the rest of the Netdata Agent. You can read more about `memory mode` in the [documentation](https://github.com/netdata/netdata/blob/master/database/README.md).
-
-The default behavior is to use the same settings as the rest of the Netdata Agent. If you wish to change them, edit the following settings:
-- `private charts memory mode`
-- `private charts history`
-
-### Optimize private metric charts storage
-
-For optimization reasons, Netdata imposes a hard limit on private metric charts. The limit is set via the `max private charts hard limit` setting (which defaults to 1000 charts). Metrics above this hard limit are still collected, but they can only be used in synthetic charts (once a metric is added to chart, it will be sent to backend servers too).
-
-If you have many ephemeral metrics collected (i.e. that you collect values for a certain amount of time), you can set the configuration option `set charts as obsolete after secs`. Setting a value in seconds here, means that Netdata will mark those metrics (and their private charts) as obsolete after the specified time has passed since the last sent metric value. Those charts will later be deleted according to the setting in `cleanup obsolete charts after secs`. Setting `set charts as obsolete after secs` to 0 (which is also the default value) will disable this functionality.
-
-Example private charts (automatically generated without any configuration):
-
-#### Counters
-
-- Scope: **count the events of something** (e.g. number of file downloads)
-- Format: `name:INTEGER|c` or `name:INTEGER|C` or `name|c`
-- StatsD increments the counter by the `INTEGER` number supplied (positive, or negative).
-
-![image](https://cloud.githubusercontent.com/assets/2662304/26131553/4a26d19c-3aa3-11e7-94e8-c53b5ed6ebc3.png)
-
-#### Gauges
-
-- Scope: **report the value of something** (e.g. cache memory used by the application server)
-- Format: `name:FLOAT|g`
-- StatsD remembers the last value supplied, and can increment or decrement the latest value if `FLOAT` begins with `+` or `-`.
-
-![image](https://cloud.githubusercontent.com/assets/2662304/26131575/5d54e6f0-3aa3-11e7-9099-bc4440cd4592.png)
-
-#### histograms
-
-- Scope: **statistics on a size of events** (e.g. statistics on the sizes of files downloaded)
-- Format: `name:FLOAT|h`
-- StatsD maintains a list of all the values supplied and provides statistics on them.
-
-![image](https://cloud.githubusercontent.com/assets/2662304/26131587/704de72a-3aa3-11e7-9ea9-0d2bb778c150.png)
-
-The same chart with `sum` unselected, to show the detail of the dimensions supported:
-![image](https://cloud.githubusercontent.com/assets/2662304/26131598/8076443a-3aa3-11e7-9ffa-ea535aee9c9f.png)
-
-#### Meters
-
-This is identical to `counter`.
-
-- Scope: **count the events of something** (e.g. number of file downloads)
-- Format: `name:INTEGER|m` or `name|m` or just `name`
-- StatsD increments the counter by the `INTEGER` number supplied (positive, or negative).
-
-![image](https://cloud.githubusercontent.com/assets/2662304/26131605/8fdf5a06-3aa3-11e7-963f-7ecf207d1dbc.png)
-
-#### Sets
-
-- Scope: **count the unique occurrences of something** (e.g. unique filenames downloaded, or unique users that downloaded files)
-- Format: `name:TEXT|s`
-- StatsD maintains a unique index of all values supplied, and reports the unique entries in it.
-
-![image](https://cloud.githubusercontent.com/assets/2662304/26131612/9eaa7b1a-3aa3-11e7-903b-d881e9a35be2.png)
-
-#### Timers
-
-- Scope: **statistics on the duration of events** (e.g. statistics for the duration of file downloads)
-- Format: `name:FLOAT|ms`
-- StatsD maintains a list of all the values supplied and provides statistics on them.
-
-![image](https://cloud.githubusercontent.com/assets/2662304/26131629/bc34f2d2-3aa3-11e7-8a07-f2fc94ba4352.png)
-
-### Synthetic StatsD charts
-
-Use synthetic charts to create dedicated sections on the dashboard to render your StatsD charts.
-
-Synthetic charts are organized in
-
-- **application** aka section in Netdata Dashboard.
-- **charts for each application** aka family in Netdata Dashboard.
-- **StatsD metrics for each chart** /aka charts and context Netdata Dashboard.
-
-> You can read more about how the Netdata Agent organizes information in the relevant [documentation](https://github.com/netdata/netdata/blob/master/web/README.md)
-
-For each application you need to create a `.conf` file in `/etc/netdata/statsd.d`.
-
-For example, if you want to monitor the application `myapp` using StatsD and Netdata, create the file `/etc/netdata/statsd.d/myapp.conf`, with this content:
-```
-[app]
- name = myapp
- metrics = myapp.*
- private charts = no
- gaps when not collected = no
- history = 60
-
-[dictionary]
- m1 = metric1
- m2 = metric2
-
-# replace 'mychart' with the chart id
-# the chart will be named: myapp.mychart
-[mychart]
- name = mychart
- title = my chart title
- family = my family
- context = chart.context
- units = tests/s
- priority = 91000
- type = area
- dimension = myapp.metric1 m1
- dimension = myapp.metric2 m2
-```
-
-Using the above configuration `myapp` should get its own section on the dashboard, having one chart with 2 dimensions.
-
-`[app]` starts a new application definition. The supported settings in this section are:
-
-- `name` defines the name of the app.
-- `metrics` is a Netdata [simple pattern](https://github.com/netdata/netdata/blob/master/libnetdata/simple_pattern/README.md). This pattern should match all the possible StatsD metrics that will be participating in the application `myapp`.
-- `private charts = yes|no`, enables or disables private charts for the metrics matched.
-- `gaps when not collected = yes|no`, enables or disables gaps on the charts of the application in case that no metrics are collected.
-- `memory mode` sets the memory mode for all charts of the application. The default is the global default for Netdata (not the global default for StatsD private charts). We suggest not to use this (we have commented it out in the example) and let your app use the global default for Netdata, which is our dbengine.
-
-- `history` sets the size of the round-robin database for this application. The default is the global default for Netdata (not the global default for StatsD private charts). This is only relevant if you use `memory mode = save`. Read more on our [metrics storage(]/docs/store/change-metrics-storage.md) doc.
-
-`[dictionary]` defines name-value associations. These are used to renaming metrics, when added to synthetic charts. Metric names are also defined at each `dimension` line. However, using the dictionary dimension names can be declared globally, for each app and is the only way to rename dimensions when using patterns. Of course the dictionary can be empty or missing.
-
-Then, add any number of charts. Each chart should start with `[id]`. The chart will be called `app_name.id`. `family` controls the submenu on the dashboard. `context` controls the alert templates. `priority` controls the ordering of the charts on the dashboard. The rest of the settings are informational.
-
-Add any number of metrics to a chart, using `dimension` lines. These lines accept 5 space separated parameters:
-
-1. the metric name, as it is collected (it has to be matched by the `metrics =` pattern of the app)
-2. the dimension name, as it should be shown on the chart
-3. an optional selector (type) of the value to shown (see below)
-4. an optional multiplier
-5. an optional divider
-6. optional flags, space separated and enclosed in quotes. All the external plugins `DIMENSION` flags can be used. Currently, the only usable flag is `hidden`, to add the dimension, but not show it on the dashboard. This is usually needed to have the values available for percentage calculation, or use them in alerts.
-
-So, the format is this:
-
-```
-dimension = [pattern] METRIC NAME TYPE MULTIPLIER DIVIDER OPTIONS
-```
-
-`pattern` is a keyword. When set, `METRIC` is expected to be a Netdata [simple pattern](https://github.com/netdata/netdata/blob/master/libnetdata/simple_pattern/README.md) that will be used to match all the StatsD metrics to be added to the chart. So, `pattern` automatically matches any number of StatsD metrics, all of which will be added as separate chart dimensions.
-
-`TYPE`, `MULTIPLIER`, `DIVIDER` and `OPTIONS` are optional.
-
-`TYPE` can be:
-
-- `events` to show the number of events received by StatsD for this metric
-- `last` to show the last value, as calculated at the flush interval of the metric (the default)
-
-Then for histograms and timers the following types are also supported:
-
-- `min`, show the minimum value
-- `max`, show the maximum value
-- `sum`, show the sum of all values
-- `average` (same as `last`)
-- `percentile`, show the 95th percentile (or any other percentile, as configured at StatsD global config)
-- `median`, show the median of all values (i.e. sort all values and get the middle value)
-- `stddev`, show the standard deviation of the values
-
-#### Example synthetic charts
-
-StatsD metrics: `foo` and `bar`.
-
-Contents of file `/etc/netdata/stats.d/foobar.conf`:
-
-```
-[app]
- name = foobarapp
- metrics = foo bar
- private charts = yes
-
-[foobar_chart1]
- title = Hey, foo and bar together
- family = foobar_family
- context = foobarapp.foobars
- units = foobars
- type = area
- dimension = foo 'foo me' last 1 1
- dimension = bar 'bar me' last 1 1
-```
-
-Metrics sent to statsd: `foo:10|g` and `bar:20|g`.
-
-Private charts:
-
-![screenshot from 2017-08-03 23-28-19](https://user-images.githubusercontent.com/2662304/28942295-7c3a73a8-78a3-11e7-88e5-a9a006bb7465.png)
-
-Synthetic chart:
-
-![screenshot from 2017-08-03 23-29-14](https://user-images.githubusercontent.com/2662304/28942317-958a2c68-78a3-11e7-853f-32850141dd36.png)
-
-#### Renaming StatsD synthetic charts' metrics
-
-You can define a dictionary to rename metrics sent by StatsD clients. This enables you to send response `"200"` and Netdata visualize it as `succesful connection`
-
-The `[dictionary]` section accepts any number of `name = value` pairs.
-
-Netdata uses this dictionary as follows:
-
-1. When a `dimension` has a non-empty `NAME`, that name is looked up at the dictionary.
-
-2. If the above lookup gives nothing, or the `dimension` has an empty `NAME`, the original StatsD metric name is looked up at the dictionary.
-
-3. If any of the above succeeds, Netdata uses the `value` of the dictionary, to set the name of the dimension. The dimensions will have as ID the original StatsD metric name, and as name, the dictionary value.
-
-Use the dictionary in 2 ways:
-
-1. set `dimension = myapp.metric1 ''` and have at the dictionary `myapp.metric1 = metric1 name`
-2. set `dimension = myapp.metric1 'm1'` and have at the dictionary `m1 = metric1 name`
-
-In both cases, the dimension will be added with ID `myapp.metric1` and will be named `metric1 name`. So, in alerts use either of the 2 as `${myapp.metric1}` or `${metric1 name}`.
-
-> keep in mind that if you add multiple times the same StatsD metric to a chart, Netdata will append `TYPE` to the dimension ID, so `myapp.metric1` will be added as `myapp.metric1_last` or `myapp.metric1_events`, etc. If you add multiple times the same metric with the same `TYPE` to a chart, Netdata will also append an incremental counter to the dimension ID, i.e. `myapp.metric1_last1`, `myapp.metric1_last2`, etc.
-
-#### Dimension patterns
-
-Netdata allows adding multiple dimensions to a chart, by matching the StatsD metrics with a Netdata simple pattern.
-
-Assume we have an API that provides StatsD metrics for each response code per method it supports, like these:
-
-```
-myapp.api.get.200
-myapp.api.get.400
-myapp.api.get.500
-myapp.api.del.200
-myapp.api.del.400
-myapp.api.del.500
-myapp.api.post.200
-myapp.api.post.400
-myapp.api.post.500
-myapp.api.all.200
-myapp.api.all.400
-myapp.api.all.500
-```
-
-In order to add all the response codes of `myapp.api.get` to a chart, we simply make the following configuration:
-
-```
-[api_get_responses]
- ...
- dimension = pattern 'myapp.api.get.* '' last 1 1
-```
-
-The above will add dimension named `200`, `400` and `500`. Netdata extracts the wildcard part of the metric name - so the dimensions will be named with whatever the `*` matched.
-
-You can rename the dimensions with this:
-
-```
-[dictionary]
- get.200 = 200 ok
- get.400 = 400 bad request
- get.500 = 500 cannot connect to db
-
-[api_get_responses]
- ...
- dimension = pattern 'myapp.api.get.* 'get.' last 1 1
-```
-
-Note that we added a `NAME` to the dimension line with `get.`. This is prefixed to the wildcarded part of the metric name, to compose the key for looking up the dictionary. So `500` became `get.500` which was looked up to the dictionary to find value `500 cannot connect to db`. This way we can have different dimension names, for each of the API methods (i.e. `get.500 = 500 cannot connect to db` while `post.500 = 500 cannot write to disk`).
-
-To add all 200s across all API methods to a chart, you can do this:
-
-```
-[ok_by_method]
- ...
- dimension = pattern 'myapp.api.*.200 '' last 1 1
-```
-
-The above will add `get`, `post`, `del` and `all` to the chart.
-
-If `all` is not wanted (a `stacked` chart does not need the `all` dimension, since the sum of the dimensions provides the total), the line should be:
-
-```
-[ok_by_method]
- ...
- dimension = pattern '!myapp.api.all.* myapp.api.*.200 '' last 1 1
-```
-
-With the above, all methods except `all` will be added to the chart.
-
-To automatically rename the methods, you can use this:
-
-```
-[dictionary]
- method.get = GET
- method.post = ADD
- method.del = DELETE
-
-[ok_by_method]
- ...
- dimension = pattern '!myapp.api.all.* myapp.api.*.200 'method.' last 1 1
-```
-
-Using the above, the dimensions will be added as `GET`, `ADD` and `DELETE`.
-
-## StatsD examples
-
-### Python
-
-It's really easy to instrument your python application with StatsD, for example using [jsocol/pystatsd](https://github.com/jsocol/pystatsd).
-
-```python
-import statsd
-c = statsd.StatsClient('localhost', 8125)
-c.incr('foo') # Increment the 'foo' counter.
-for i in range(100000000):
- c.incr('bar')
- c.incr('foo')
- if i % 3:
- c.decr('bar')
- c.timing('stats.timed', 320) # Record a 320ms 'stats.timed'.
-```
-
-You can find detailed documentation in their [documentation page](https://statsd.readthedocs.io/en/v3.3/).
-
-### Javascript and Node.js
-
-Using the client library by [sivy/node-statsd](https://github.com/sivy/node-statsd), you can easily embed StatsD into your Node.js project.
-
-```javascript
- var StatsD = require('node-statsd'),
- client = new StatsD();
-
- // Timing: sends a timing command with the specified milliseconds
- client.timing('response_time', 42);
-
- // Increment: Increments a stat by a value (default is 1)
- client.increment('my_counter');
-
- // Decrement: Decrements a stat by a value (default is -1)
- client.decrement('my_counter');
-
- // Using the callback
- client.set(['foo', 'bar'], 42, function(error, bytes){
- //this only gets called once after all messages have been sent
- if(error){
- console.error('Oh noes! There was an error:', error);
- } else {
- console.log('Successfully sent', bytes, 'bytes');
- }
- });
-
- // Sampling, tags and callback are optional and could be used in any combination
- client.histogram('my_histogram', 42, 0.25); // 25% Sample Rate
- client.histogram('my_histogram', 42, ['tag']); // User-defined tag
- client.histogram('my_histogram', 42, next); // Callback
- client.histogram('my_histogram', 42, 0.25, ['tag']);
- client.histogram('my_histogram', 42, 0.25, next);
- client.histogram('my_histogram', 42, ['tag'], next);
- client.histogram('my_histogram', 42, 0.25, ['tag'], next);
-```
-### Other languages
-
-You can also use StatsD with:
-- Golang, thanks to [alexcesaro/statsd](https://github.com/alexcesaro/statsd)
-- Ruby, thanks to [reinh/statsd](https://github.com/reinh/statsd)
-- Java, thanks to [DataDog/java-dogstatsd-client](https://github.com/DataDog/java-dogstatsd-client)
-
-
-### Shell
-
-Getting the proper support for a programming language is not always easy, but the Unix shell is available on most Unix systems. You can use shell and `nc` to instrument your systems and send metric data to Netdata's StatsD implementation.
-
-Using the method you can send metrics from any script. You can generate events like: backup.started, backup.ended, backup.time, or even tail logs and convert them to metrics.
-
-> **IMPORTANT**:
->
-> To send StatsD messages you need from the `netcat` package, the `nc` command.
-> There are multiple versions of this package. Please try to experiment with the `nc` command you have available on your right system, to find the right parameters.
->
-> In the examples below, we assume the `openbsd-netcat` is installed.
-
-If you plan to send short StatsD events at sporadic occasions, use UDP. The messages should not be too long (remember, most networks support up to 1500 bytes MTU, which is also the limit for StatsD messages over UDP). The good thing is that using UDP will not block your script, even if the StatsD server is not there (UDP messages are "fire-and-forget").
-
-
-For UDP use this:
-
-```sh
-echo "APPLICATION.METRIC:VALUE|TYPE" | nc -u -w 0 localhost 8125
-```
-
-`-u` turns on UDP, `-w 0` tells `nc` not to wait for a response from StatsD (idle time to close the connection).
-
-where:
-
-- `APPLICATION` is any name for your application
-- `METRIC` is the name for the specific metric
-- `VALUE` is the value for that metric (**meters**, **counters**, **gauges**, **timers** and **histograms** accept integer/decimal/fractional numbers, **sets** and **dictionaries** accept strings)
-- `TYPE` is one of `m`, `c`, `g`, `ms`, `h`, `s`, `d` to define the metric type.
-
-For tailing a log and converting it to metrics, do something like this:
-
-```sh
-tail -f some.log | awk 'awk commands to parse the log and format statsd metrics' | nc -N -w 120 localhost 8125
-```
-
-`-N` tells `nc` to close the socket once it receives EOF on its input. `-w 120` tells `nc` to stop if the connection is idle for 120 seconds. The timeout is needed to stop the `nc` command if you restart Netdata while `nc` is connected to it. Without it, `nc` will sit idle forever.
-
-When you embed the above commands to a script, you may notice that all the metrics are sent to StatsD with a delay. They are buffered in the pipes `|`. You can turn them to real-time by prepending each command with `stdbuf -i0 -oL -eL command to be run`, like this:
-
-```sh
-stdbuf -i0 -oL -eL tail -f some.log |\
- stdbuf -i0 -oL -eL awk 'awk commands to parse the log and format statsd metrics' |\
- stdbuf -i0 -oL -eL nc -N -w 120 localhost 8125
-```
-
-If you use `mawk` you also need to run awk with `-W interactive`.
-
-Examples:
-
-To set `myapp.used_memory` as gauge to value `123456`, use:
-
-```sh
-echo "myapp.used_memory:123456|g|#units:bytes" | nc -u -w 0 localhost 8125
-```
-
-To increment `myapp.files_sent` by `10`, as a counter, use:
-
-```sh
-echo "myapp.files_sent:10|c|#units:files" | nc -u -w 0 localhost 8125
-```
-
-You can send multiple metrics like this:
-
-```sh
-# send multiple metrics via UDP
-printf "myapp.used_memory:123456|g|#units:bytes\nmyapp.files_sent:10|c|#units:files\n" | nc -u -w 0 localhost 8125
-```
-
-Remember, for UDP communication each packet should not exceed the MTU. So, if you plan to push too many metrics at once, prefer TCP communication:
-
-```sh
-# send multiple metrics via TCP
-cat /tmp/statsd.metrics.txt | nc -N -w 120 localhost 8125
-```
-
-You can also use this little function to take care of all the details:
-
-```sh
-#!/usr/bin/env bash
-
-# we assume nc is from the openbsd-netcat package
-
-STATSD_HOST="localhost"
-STATSD_PORT="8125"
-statsd() {
- local options="-u -w 0" all="${*}"
-
- # replace all spaces with newlines
- all="${all// /\\n}"
-
- # if the string length of all parameters given is above 1000, use TCP
- [ "${#all}" -gt 1000 ] && options="-N -w 0"
-
- # send the metrics to statsd
- printf "${all}\n" | nc ${options} ${STATSD_HOST} ${STATSD_PORT} || return 1
-
- return 0
-}
-
-if [ ! -z "${*}" ]
-then
- statsd "${@}"
-fi
-```
-
-You can use it like this:
-
-```sh
-# first, source it in your script
-source statsd.sh
-
-# then, at any point:
-statsd "myapp.used_memory:123456|g|#units:bytes" "myapp.files_sent:10|c|#units:files" ...
-```
-
-or even at a terminal prompt, like this:
-
-```sh
-./statsd.sh "myapp.used_memory:123456|g|#units:bytes" "myapp.files_sent:10|c|#units:files" ...
-```
-
-The function is smart enough to call `nc` just once and pass all the metrics to it. It will also automatically switch to TCP if the metrics to send are above 1000 bytes.
-
-If you have gotten thus far, make sure to check out our [community forums](https://community.netdata.cloud) to share your experience using Netdata with StatsD.
-
-## StatsD Step By Step Guide
-
-In this guide, we'll go through a scenario of visualizing our data in Netdata in a matter of seconds using
-[k6](https://k6.io), an open-source tool for automating load testing that outputs metrics to the StatsD format.
-
-Although we'll use k6 as the use-case, the same principles can be applied to every application that supports
-the StatsD protocol. Simply enable the StatsD output and point it to the node that runs Netdata, which is `localhost` in this case.
-
-In general, the process for creating a StatsD collector can be summarized in 2 steps:
-
-- Run an experiment by sending StatsD metrics to Netdata, without any prior configuration. This will create
- a chart per metric (called private charts) and will help you verify that everything works as expected from the application side of things.
-
- - Make sure to reload the dashboard tab **after** you start sending data to Netdata.
-
-- Create a configuration file for your app using [edit-config](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md): `sudo ./edit-config
- statsd.d/myapp.conf`
-
- - Each app will have it's own section in the right-hand menu.
-
-Now, let's see the above process in detail.
-
-### Prerequisites
-
-- A node with the [Netdata](https://github.com/netdata/netdata/blob/master/packaging/installer/README.md) installed.
-- An application to instrument. For this guide, that will be [k6](https://k6.io/docs/getting-started/installation).
-
-### Understanding the metrics
-
-The real in instrumenting an application with StatsD for you is to decide what metrics you
-want to visualize and how you want them grouped. In other words, you need decide which metrics
-will be grouped in the same charts and how the charts will be grouped on Netdata's dashboard.
-
-Start with documentation for the particular application that you want to monitor (or the
-technological stack that you are using). In our case, the
-[k6 documentation](https://k6.io/docs/using-k6/metrics/) has a whole page dedicated to the
-metrics output by k6, along with descriptions.
-
-If you are using StatsD to monitor an existing application, you don't have much control over
-these metrics. For example, k6 has a type called `trend`, which is identical to timers and histograms.
-Thus, _k6 is clearly dictating_ which metrics can be used as histograms and simple gauges.
-
-On the other hand, if you are instrumenting your own code, you will need to not only decide what are
-the "things" that you want to measure, but also decide which StatsD metric type is the appropriate for each.
-
-### Use private charts to see all available metrics
-
-In Netdata, every metric will receive its own chart, called a `private chart`. Although in the
-final implementation this is something that we will disable, since it can create considerable noise
-(imagine having 100s of metrics), it’s very handy while building the configuration file.
-
-You can get a quick visual representation of the metrics and their type (e.g it’s a gauge, a timer, etc.).
-
-An important thing to notice is that StatsD has different types of metrics, as illustrated in the
-[supported metrics](#metrics-supported-by-netdata). Histograms and timers support mathematical operations
-to be performed on top of the baseline metric, like reporting the `average` of the value.
-
-Here are some examples of default private charts. You can see that the histogram private charts will
-visualize all the available operations.
-
-**Gauge private chart**
-
-![Gauge metric example](https://i.imgur.com/Sr5nJEV.png)
-
-**Histogram private chart**
-
-![Timer metric example](https://i.imgur.com/P4p0hvq.png)
-
-### Create a new StatsD configuration file
-
-Start by creating a new configuration file under the `statsd.d/` folder in the
-[Netdata config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
-Use [`edit-config`](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#use-edit-config-to-edit-configuration-files)
-to create a new file called `k6.conf`.
-
-```bash=
-sudo ./edit-config statsd.d/k6.conf
-```
-
-Copy the following configuration into your file as a starting point.
-
-```conf
-[app]
- name = k6
- metrics = k6*
- private charts = yes
- gaps when not collected = no
- memory mode = dbengine
-```
-
-Next, you need is to understand how to organize metrics in Netdata’s StatsD.
-
-#### Synthetic charts
-
-Netdata lets you group the metrics exposed by your instrumented application with _synthetic charts_.
-
-First, create a `[dictionary]` section to transform the names of the metrics into human-readable equivalents.
-`http_req_blocked`, `http_req_connecting`, `http_req_receiving`, and `http_reqs` are all metrics exposed by k6.
-
-```
-[dictionary]
- http_req_blocked = Blocked HTTP Requests
- http_req_connecting = Connecting HTTP Requests
- http_req_receiving = Receiving HTTP Requests
- http_reqs = Total HTTP requests
-```
-
-Continue this dictionary process with any other metrics you want to collect with Netdata.
-
-#### Families and context
-
-Families and context are additional ways to group metrics. Families control the submenu at right-hand menu and
-it's a subcategory of the section. Given the metrics given by K6, we are organizing them in 2 major groups,
-or `families`: `k6 native metrics` and `http metrics`.
-
-Context is a second way to group metrics, when the metrics are of the same nature but different origin. In
-our case, if we ran several different load testing experiments side-by-side, we could define the same app,
-but different context (e.g `http_requests.experiment1`, `http_requests.experiment2`).
-
-Find more details about family and context in our [documentation](https://github.com/netdata/netdata/blob/master/web/README.md#families).
-
-#### Dimensions
-
-Now, having decided on how we are going to group the charts, we need to define how we are going to group
-metrics into different charts. This is particularly important, since we decide:
-
-- What metrics **not** to show, since they are not useful for our use-case.
-- What metrics to consolidate into the same charts, so as to reduce noise and increase visual correlation.
-
-The dimension option has this syntax: `dimension = [pattern] METRIC NAME TYPE MULTIPLIER DIVIDER OPTIONS`
-
-- **pattern**: A keyword that tells the StatsD server the `METRIC` string is actually a
- [simple pattern](https://github.com/netdata/netdata/blob/master/libnetdata/simple_pattern/README.md).
- We don't use simple patterns in the example, but if we wanted to visualize all the `http_req` metrics, we
- could have a single dimension: `dimension = pattern 'k6.http_req*' last 1 1`. Find detailed examples with
- patterns in [dimension patterns](https://github.com/netdata/netdata/blob/master/collectors/statsd.plugin/README.md#dimension-patterns).
-
-- **METRIC** The id of the metric as it comes from the client. You can easily find this in the private charts above,
- for example: `k6.http_req_connecting`.
-
-- **NAME**: The name of the dimension. You can use the dictionary to expand this to something more human-readable.
-
-- **TYPE**:
-
- - For all charts:
- - `events`: The number of events (data points) received by the StatsD server
- - `last`: The last value that the server received
-
- - For histograms and timers:
- - `min`, `max`, `sum`, `average`, `percentile`, `median`, `stddev`: This is helpful if you want to see
- different representations of the same value. You can find an example at the `[iteration_duration]`
- above. Note that the baseline `metric` is the same, but the `name` of the dimension is different,
- since we use the baseline, but we perform a computation on it, creating a different final metric for
- visualization(dimension).
-
-- **MULTIPLIER DIVIDER**: Handy if you want to convert Kilobytes to Megabytes or you want to give negative value.
- The second is handy for better visualization of send/receive. You can find an example at the **packets** submenu of the **IPv4 Networking Section**.
-
-If you define a chart, run Netdata to visualize metrics, and then add or remove a dimension from that chart,
-this will result in a new chart with the same name, confusing Netdata. If you change the dimensions of the chart,
-make sure to also change the `name` of that chart, since it serves as the `id` of that chart in Netdata's storage.
-(e.g http_req --> http_req_1).
-
-#### Finalize your StatsD configuration file
-
-It's time to assemble all the pieces together and create the synthetic charts that will consist our application
-dashboard in Netdata. We can do it in a few simple steps:
-
-- Decide which metrics we want to use (we have viewed all of them as private charts). For example, we want to use
- `k6.http_requests`, `k6.vus`, etc.
-
-- Decide how we want organize them in different synthetic charts. For example, we want `k6.http_requests`, `k6.vus`
- on their own, but `k6.http_req_blocked` and `k6.http_req_connecting` on the same chart.
-
-- For each synthetic chart, we define a **unique** name and a human readable title.
-
-- We decide at which `family` (submenu section) we want each synthetic chart to belong to. For example, here we
- have defined 2 families: `http requests`, `k6_metrics`.
-
-- If we have multiple instances of the same metric, we can define different contexts, (Optional).
-
-- We define a dimension according to the syntax we highlighted above.
-
-- We define a type for each synthetic chart (line, area, stacked)
-
-- We define the units for each synthetic chart.
-
-Following the above steps, we append to the `k6.conf` that we defined above, the following configuration:
-
-```
-[http_req_total]
- name = http_req_total
- title = Total HTTP Requests
- family = http requests
- context = k6.http_requests
- dimension = k6.http_reqs http_reqs last 1 1 sum
- type = line
- units = requests/s
-
-[vus]
- name = vus
- title = Virtual Active Users
- family = k6_metrics
- dimension = k6.vus vus last 1 1
- dimension = k6.vus_max vus_max last 1 1
- type = line
- unit = vus
-
-[iteration_duration]
- name = iteration_duration_2
- title = Iteration duration
- family = k6_metrics
- dimension = k6.iteration_duration iteration_duration last 1 1
- dimension = k6.iteration_duration iteration_duration_max max 1 1
- dimension = k6.iteration_duration iteration_duration_min min 1 1
- dimension = k6.iteration_duration iteration_duration_avg avg 1 1
- type = line
- unit = s
-
-[dropped_iterations]
- name = dropped_iterations
- title = Dropped Iterations
- family = k6_metrics
- dimension = k6.dropped_iterations dropped_iterations last 1 1
- units = iterations
- type = line
-
-[data]
- name = data
- title = K6 Data
- family = k6_metrics
- dimension = k6.data_received data_received last 1 1
- dimension = k6.data_sent data_sent last -1 1
- units = kb/s
- type = area
-
-[http_req_status]
- name = http_req_status
- title = HTTP Requests Status
- family = http requests
- dimension = k6.http_req_blocked http_req_blocked last 1 1
- dimension = k6.http_req_connecting http_req_connecting last 1 1
- units = ms
- type = line
-
-[http_req_duration]
- name = http_req_duration
- title = HTTP requests duration
- family = http requests
- dimension = k6.http_req_sending http_req_sending last 1 1
- dimension = k6.http_req_waiting http_req_waiting last 1 1
- dimension = k6.http_req_receiving http_req_receiving last 1 1
- units = ms
- type = stacked
-```
-
-Note that Netdata will report the rate for metrics and counters, even if k6 or another application
-sends an _absolute_ number. For example, k6 sends absolute HTTP requests with `http_reqs`,
-but Netdata visualizes that in `requests/second`.
-
-To enable this StatsD configuration, [restart Netdata](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md).
-
-### Final touches
-
-At this point, you have used StatsD to gather metrics for k6, creating a whole new section in your
-Netdata dashboard in the process. Moreover, you can further customize the icon of the particular section,
-as well as the description for each chart.
-
-To edit the section, please follow the Netdata [documentation](https://github.com/netdata/netdata/blob/master/web/gui/README.md#customizing-the-local-dashboard).
-
-While the following configuration will be placed in a new file, as the documentation suggests, it is
-instructing to use `dashboard_info.js` as a template. Open the file and see how the rest of sections and collectors have been defined.
-
-```javascript=
-netdataDashboard.menu = {
- 'k6': {
- title: 'K6 Load Testing',
- icon: '<i class="fas fa-cogs"></i>',
- info: 'k6 is an open-source load testing tool and cloud service providing the best developer experience for API performance testing.'
- },
- .
- .
- .
-```
-
-We can then add a description for each chart. Simply find the following section in `dashboard_info.js` to understand how a chart definitions are used:
-
-```javascript=
-netdataDashboard.context = {
- 'system.cpu': {
- info: function (os) {
- void (os);
- return 'Total CPU utilization (all cores). 100% here means there is no CPU idle time at all. You can get per core usage at the <a href="#menu_cpu">CPUs</a> section and per application usage at the <a href="#menu_apps">Applications Monitoring</a> section.'
- + netdataDashboard.sparkline('<br/>Keep an eye on <b>iowait</b> ', 'system.cpu', 'iowait', '%', '. If it is constantly high, your disks are a bottleneck and they slow your system down.')
- + netdataDashboard.sparkline('<br/>An important metric worth monitoring, is <b>softirq</b> ', 'system.cpu', 'softirq', '%', '. A constantly high percentage of softirq may indicate network driver issues.');
- },
- valueRange: "[0, 100]"
- },
-```
-
-Afterwards, you can open your `custom_dashboard_info.js`, as suggested in the documentation linked above,
-and add something like the following example:
-
-```javascript=
-netdataDashboard.context = {
- 'k6.http_req_duration': {
- info: "Total time for the request. It's equal to http_req_sending + http_req_waiting + http_req_receiving (i.e. how long did the remote server take to process the request and respond, without the initial DNS lookup/connection times)"
- },
-
-```
-The chart is identified as ``<section_name>.<chart_name>``.
-
-These descriptions can greatly help the Netdata user who is monitoring your application in the midst of an incident.
-
-The `info` field supports `html`, embedding useful links and instructions in the description.
-
-### Vendoring a new collector
-
-While we learned how to visualize any data source in Netdata using the StatsD protocol, we have also created a new collector.
-
-As long as you use the same underlying collector, every new `myapp.conf` file will create a new data
-source and dashboard section for Netdata. Netdata loads all the configuration files by default, but it will
-**not** create dashboard sections or charts, unless it starts receiving data for that particular data source.
-This means that we can now share our collector with the rest of the Netdata community.
-
-- Make sure you follow the [contributing guide](https://github.com/netdata/.github/edit/main/CONTRIBUTING.md)
-- Fork the netdata/netdata repository
-- Place the configuration file inside `netdata/collectors/statsd.plugin`
-- Add a reference in `netdata/collectors/statsd.plugin/Makefile.am`. For example, if we contribute the `k6.conf` file:
-```Makefile
-dist_statsdconfig_DATA = \
- example.conf \
- k6.conf \
- $(NULL)
-```
-
-
diff --git a/collectors/systemd-journal.plugin/Makefile.am b/collectors/systemd-journal.plugin/Makefile.am
deleted file mode 100644
index 48f667c1b..000000000
--- a/collectors/systemd-journal.plugin/Makefile.am
+++ /dev/null
@@ -1,16 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- systemd-journal-self-signed-certs.sh \
- forward_secure_sealing.md \
- active_journal_centralization_guide_no_encryption.md \
- passive_journal_centralization_guide_no_encryption.md \
- passive_journal_centralization_guide_self_signed_certs.md \
- $(NULL)
-
-dist_libconfig_DATA = \
- $(NULL)
diff --git a/collectors/systemd-journal.plugin/README.md b/collectors/systemd-journal.plugin/README.md
deleted file mode 100644
index c3c639045..000000000
--- a/collectors/systemd-journal.plugin/README.md
+++ /dev/null
@@ -1,472 +0,0 @@
-
-# `systemd` journal plugin
-
-[KEY FEATURES](#key-features) | [JOURNAL SOURCES](#journal-sources) | [JOURNAL FIELDS](#journal-fields) |
-[PLAY MODE](#play-mode) | [FULL TEXT SEARCH](#full-text-search) | [PERFORMANCE](#query-performance) |
-[CONFIGURATION](#configuration-and-maintenance) | [FAQ](#faq)
-
-The `systemd` journal plugin by Netdata makes viewing, exploring and analyzing `systemd` journal logs simple and
-efficient.
-It automatically discovers available journal sources, allows advanced filtering, offers interactive visual
-representations and supports exploring the logs of both individual servers and the logs on infrastructure wide
-journal centralization servers.
-
-![image](https://github.com/netdata/netdata/assets/2662304/691b7470-ec56-430c-8b81-0c9e49012679)
-
-## Key features
-
-- Works on both **individual servers** and **journal centralization servers**.
-- Supports `persistent` and `volatile` journals.
-- Supports `system`, `user`, `namespaces` and `remote` journals.
-- Allows filtering on **any journal field** or **field value**, for any time-frame.
-- Allows **full text search** (`grep`) on all journal fields, for any time-frame.
-- Provides a **histogram** for log entries over time, with a break down per field-value, for any field and any
- time-frame.
-- Works directly on journal files, without any other third-party components.
-- Supports coloring log entries, the same way `journalctl` does.
-- In PLAY mode provides the same experience as `journalctl -f`, showing new log entries immediately after they are
- received.
-
-### Prerequisites
-
-`systemd-journal.plugin` is a Netdata Function Plugin.
-
-To protect your privacy, as with all Netdata Functions, a free Netdata Cloud user account is required to access it.
-For more information check [this discussion](https://github.com/netdata/netdata/discussions/16136).
-
-### Limitations
-
-#### Plugin availability
-
-The following are limitations related to the availability of the plugin:
-
-- Netdata versions prior to 1.44 shipped in a docker container do not include this plugin.
- The problem is that `libsystemd` is not available in Alpine Linux (there is a `libsystemd`, but it is a dummy that
- returns failure on all calls). Starting with Netdata version 1.44, Netdata containers use a Debian base image
- making this plugin available when Netdata is running in a container.
-- For the same reason (lack of `systemd` support for Alpine Linux), the plugin is not available on `static` builds of
- Netdata (which are based on `muslc`, not `glibc`). If your Netdata is installed in `/opt/netdata` you most likely have
- a static build of Netdata.
-- On old systemd systems (like Centos 7), the plugin runs always in "full data query" mode, which makes it slower. The
- reason, is that systemd API is missing some important calls we need to use the field indexes of `systemd` journal.
- However, when running in this mode, the plugin offers also negative matches on the data (like filtering for all logs
- that do not have set some field), and this is the reason "full data query" mode is also offered as an option even on
- newer versions of `systemd`.
-
-#### `systemd` journal features
-
-The following are limitations related to the features of `systemd` journal:
-
-- This plugin assumes that binary field values are text fields with newlines in them. `systemd-journal` has the ability
- to support binary fields, without specifying the nature of the binary data. However, binary fields are commonly used
- to store log entries that include multiple lines of text. The plugin treats all binary fields are multi-line text.
-- This plugin does not support multiple values per field for any given log entry. `systemd` journal has the ability to
- accept the same field key, multiple times, with multiple values on a single log entry. This plugin will present the
- last value and ignore the others for this log entry.
-- This plugin will only read journal files located in `/var/log/journal` or `/run/log/journal`. `systemd-journal-remote` has the
- ability to store journal files anywhere (user configured). If journal files are not located in `/var/log/journal`
- or `/run/log/journal` (and any of their subdirectories), the plugin will not find them. A simple solution is to link
- the other directories somewhere inside `/var/log/journal`. The plugin will pick them up, even if a sub-directory of
- `/var/log/journal` is a link to a directory outside `/var/log/journal`.
-
-Other than the above, this plugin supports all features of `systemd` journals.
-
-## Journal Sources
-
-The plugin automatically detects the available journal sources, based on the journal files available in
-`/var/log/journal` (persistent logs) and `/run/log/journal` (volatile logs).
-
-![journal-sources](https://github.com/netdata/netdata/assets/2662304/28e63a3e-6809-4586-b3b0-80755f340e31)
-
-The plugin, by default, merges all journal sources together, to provide a unified view of all log messages available.
-
-> To improve query performance, we recommend selecting the relevant journal source, before doing more analysis on the
-> logs.
-
-### `system` journals
-
-`system` journals are the default journals available on all `systemd` based systems.
-
-`system` journals contain:
-
-- kernel log messages (via `kmsg`),
-- audit records, originating from the kernel audit subsystem,
-- messages received by `systemd-journald` via `syslog`,
-- messages received via the standard output and error of service units,
-- structured messages received via the native journal API.
-
-### `user` journals
-
-Unlike `journalctl`, the Netdata plugin allows viewing, exploring and querying the journal files of **all users**.
-
-By default, each user, with a UID outside the range of system users (0 - 999), dynamic service users,
-and the nobody user (65534), will get their own set of `user` journal files. For more information about
-this policy check [Users, Groups, UIDs and GIDs on systemd Systems](https://systemd.io/UIDS-GIDS/).
-
-Keep in mind that `user` journals are merged with the `system` journals when they are propagated to a journal
-centralization server. So, at the centralization server, the `remote` journals contain both the `system` and `user`
-journals of the sender.
-
-### `namespaces` journals
-
-The plugin auto-detects the namespaces available and provides a list of all namespaces at the "sources" list on the UI.
-
-Journal namespaces are both a mechanism for logically isolating the log stream of projects consisting
-of one or more services from the rest of the system and a mechanism for improving performance.
-
-`systemd` service units may be assigned to a specific journal namespace through the `LogNamespace=` unit file setting.
-
-Keep in mind that namespaces require special configuration to be propagated to a journal centralization server.
-This makes them a little more difficult to handle, from the administration perspective.
-
-### `remote` journals
-
-Remote journals are created by `systemd-journal-remote`. This `systemd` feature allows creating logs centralization
-points within your infrastructure, based exclusively on `systemd`.
-
-Usually `remote` journals are named by the IP of the server sending these logs. The Netdata plugin automatically
-extracts these IPs and performs a reverse DNS lookup to find their hostnames. When this is successful,
-`remote` journals are named by the hostnames of the origin servers.
-
-For information about configuring a journal centralization server,
-check [this FAQ item](#how-do-i-configure-a-journal-centralization-server).
-
-## Journal Fields
-
-`systemd` journals are designed to support multiple fields per log entry. The power of `systemd` journals is that,
-unlike other log management systems, it supports dynamic and variable fields for each log message,
-while all fields and their values are indexed for fast querying.
-
-This means that each application can log messages annotated with its own unique fields and values, and `systemd`
-journals will automatically index all of them, without any configuration or manual action.
-
-For a description of the most frequent fields found in `systemd` journals, check `man systemd.journal-fields`.
-
-Fields found in the journal files are automatically added to the UI in multiple places to help you explore
-and filter the data.
-
-The plugin automatically enriches certain fields to make them more user-friendly:
-
-- `_BOOT_ID`: the hex value is annotated with the timestamp of the first message encountered for this boot id.
-- `PRIORITY`: the numeric value is replaced with the human-readable name of each priority.
-- `SYSLOG_FACILITY`: the encoded value is replaced with the human-readable name of each facility.
-- `ERRNO`: the numeric value is annotated with the short name of each value.
-- `_UID` `_AUDIT_LOGINUID`, `_SYSTEMD_OWNER_UID`, `OBJECT_UID`, `OBJECT_SYSTEMD_OWNER_UID`, `OBJECT_AUDIT_LOGINUID`:
- the local user database is consulted to annotate them with usernames.
-- `_GID`, `OBJECT_GID`: the local group database is consulted to annotate them with group names.
-- `_CAP_EFFECTIVE`: the encoded value is annotated with a human-readable list of the linux capabilities.
-- `_SOURCE_REALTIME_TIMESTAMP`: the numeric value is annotated with human-readable datetime in UTC.
-- `MESSAGE_ID`: for the known `MESSAGE_ID`s, the value is replaced with the well known name of the event.
-
-The values of all other fields are presented as found in the journals.
-
-> IMPORTANT:
-> The UID and GID annotations are added during presentation and are taken from the server running the plugin.
-> For `remote` sources, the names presented may not reflect the actual user and group names on the origin server.
-> The numeric value will still be visible though, as-is on the origin server.
-
-The annotations are not searchable with full-text search. They are only added for the presentation of the fields.
-
-### Journal fields as columns in the table
-
-All journal fields available in the journal files are offered as columns on the UI. Use the gear button above the table:
-
-![image](https://github.com/netdata/netdata/assets/2662304/cd75fb55-6821-43d4-a2aa-033792c7f7ac)
-
-### Journal fields as additional info to each log entry
-
-When you click a log line, the `info` sidebar will open on the right of the screen, to provide the full list of fields
-related to this log line. You can close this `info` sidebar, by selecting the filter icon at its top.
-
-![image](https://github.com/netdata/netdata/assets/2662304/3207794c-a61b-444c-8ffe-6c07cbc90ae2)
-
-### Journal fields as filters
-
-The plugin presents a select list of fields as filters to the query, with counters for each of the possible values
-for the field. This list can used to quickly check which fields and values are available for the entire time-frame
-of the query.
-
-Internally the plugin has:
-
-1. A white-list of fields, to be presented as filters.
-2. A black-list of fields, to prevent them from becoming filters. This list includes fields with a very high
- cardinality, like timestamps, unique message ids, etc. This is mainly for protecting the server's performance,
- to avoid building in memory indexes for the fields that almost each of their values is unique.
-
-Keep in mind that the values presented in the filters, and their sorting is affected by the "full data queries"
-setting:
-
-![image](https://github.com/netdata/netdata/assets/2662304/ac710d46-07c2-487b-8ce3-e7f767b9ae0f)
-
-When "full data queries" is off, empty values are hidden and cannot be selected. This is due to a limitation of
-`libsystemd` that does not allow negative or empty matches. Also, values with zero counters may appear in the list.
-
-When "full data queries" is on, Netdata is applying all filtering to the data (not `libsystemd`), but this means
-that all the data of the entire time-frame, without any filtering applied, have to be read by the plugin to prepare
-the response required. So, "full data queries" can be significantly slower over long time-frames.
-
-### Journal fields as histogram sources
-
-The plugin presents a histogram of the number of log entries across time.
-
-The data source of this histogram can be any of the fields that are available as filters.
-For each of the values this field has, across the entire time-frame of the query, the histogram will get corresponding
-dimensions, showing the number of log entries, per value, over time.
-
-The granularity of the histogram is adjusted automatically to have about 150 columns visible on screen.
-
-The histogram presented by the plugin is interactive:
-
-- **Zoom**, either with the global date-time picker, or the zoom tool in the histogram's toolbox.
-- **Pan**, either with global date-time picker, or by dragging with the mouse the chart to the left or the right.
-- **Click**, to quickly jump to the highlighted point in time in the log entries.
-
-![image](https://github.com/netdata/netdata/assets/2662304/d3dcb1d1-daf4-49cf-9663-91b5b3099c2d)
-
-## PLAY mode
-
-The plugin supports PLAY mode, to continuously update the screen with new log entries found in the journal files.
-Just hit the "play" button at the top of the Netdata dashboard screen.
-
-On centralized log servers, PLAY mode provides a unified view of all the new logs encountered across the entire
-infrastructure,
-from all hosts sending logs to the central logs server via `systemd-remote`.
-
-## Full-text search
-
-The plugin supports searching for any text on all fields of the log entries.
-
-Full text search is combined with the selected filters.
-
-The text box accepts asterisks `*` as wildcards. So, `a*b*c` means match anything that contains `a`, then `b` and
-then `c` with anything between them.
-
-Spaces are treated as OR expressions. So that `a*b c*d` means `a*b OR c*d`.
-
-Negative expressions are supported, by prefixing any string with `!`. Example: `!systemd *` means match anything that
-does not contain `systemd` on any of its fields.
-
-## Query performance
-
-Journal files are designed to be accessed by multiple readers and one writer, concurrently.
-
-Readers (like this Netdata plugin), open the journal files and `libsystemd`, behind the scenes, maps regions
-of the files into memory, to satisfy each query.
-
-On logs aggregation servers, the performance of the queries depend on the following factors:
-
-1. The **number of files** involved in each query.
-
- This is why we suggest to select a source when possible.
-
-2. The **speed of the disks** hosting the journal files.
-
- Journal files perform a lot of reading while querying, so the fastest the disks, the faster the query will finish.
-
-3. The **memory available** for caching parts of the files.
-
- Increased memory will help the kernel cache the most frequently used parts of the journal files, avoiding disk I/O
- and speeding up queries.
-
-4. The **number of filters** applied.
-
- Queries are significantly faster when just a few filters are selected.
-
-In general, for a faster experience, **keep a low number of rows within the visible timeframe**.
-
-Even on long timeframes, selecting a couple of filters that will result in a **few dozen thousand** log entries
-will provide fast / rapid responses, usually less than a second. To the contrary, viewing timeframes with **millions
-of entries** may result in longer delays.
-
-The plugin aborts journal queries when your browser cancels inflight requests. This allows you to work on the UI
-while there are background queries running.
-
-At the time of this writing, this Netdata plugin is about 25-30 times faster than `journalctl` on queries that access
-multiple journal files, over long time-frames.
-
-During the development of this plugin, we submitted, to `systemd`, a number of patches to improve `journalctl`
-performance by a factor of 14:
-
-- <https://github.com/systemd/systemd/pull/29365>
-- <https://github.com/systemd/systemd/pull/29366>
-- <https://github.com/systemd/systemd/pull/29261>
-
-However, even after these patches are merged, `journalctl` will still be 2x slower than this Netdata plugin,
-on multi-journal queries.
-
-The problem lies in the way `libsystemd` handles multi-journal file queries. To overcome this problem,
-the Netdata plugin queries each file individually and it then it merges the results to be returned.
-This is transparent, thanks to the `facets` library in `libnetdata` that handles on-the-fly indexing, filtering,
-and searching of any dataset, independently of its source.
-
-## Performance at scale
-
-On busy logs servers, or when querying long timeframes that match millions of log entries, the plugin has a sampling
-algorithm to allow it respond promptly. It works like this:
-
-1. The latest 500k log entries are queried in full, evaluating all the fields of every single log entry. This evaluation
- allows counting the unique values per field, updating the counters next to each value at the filters section of the
- dashboard.
-2. When the latest 500k log entries have been processed and there are more data to read, the plugin divides evenly 500k
- more log entries to the number of journal files matched by the query. So, it will continue to evaluate all the fields
- of all log entries, up to the budget per file, aiming to fully query 1 million log entries in total.
-3. When the budget is hit for a given file, the plugin continues to scan log entries, but this time it does not evaluate
- the fields and their values, so the counters per field and value are not updated. These unsampled log entries are
- shown in the histogram with the label `[unsampled]`.
-4. The plugin continues to count `[unsampled]` entries until as many as sampled entries have been evaluated and at least
- 1% of the journal file has been processed.
-5. When the `[unsampled]` budget is exhausted, the plugin stops processing the journal file and based on the processing
- completed so far and the number of entries in the journal file, it estimates the remaining number of log entries in
- that file. This is shown as `[estimated]` at the histogram.
-6. In systemd versions 254 or later, the plugin fetches the unique sequence number of each log entry and calculates the
- the percentage of the file matched by the query, versus the total number of the log entries in the journal file.
-7. In systemd versions prior to 254, the plugin estimates the number of entries the journal file contributes to the
- query, using the amount of log entries matched it vs. the total duration the log file has entries for.
-
-The above allow the plugin to respond promptly even when the number of log entries in the journal files is several
-dozens millions, while providing accurate estimations of the log entries over time at the histogram and enough counters
-at the fields filtering section to help users get an overview of the whole timeframe.
-
-The fact that the latest 500k log entries and 1% of all journal files (which are spread over time) have been fully
-evaluated, including counting the number of appearances for each field value, the plugin usually provides an accurate
-representation of the whole timeframe.
-
-Keep in mind that although the plugin is quite effective and responds promptly when there are hundreds of journal files
-matching a query, response times may be longer when there are several thousands of smaller files. systemd versions 254+
-attempt to solve this problem by allowing `systemd-journal-remote` to create larger files. However, for systemd
-versions prior to 254, `systemd-journal-remote` creates files of up to 32MB each, which when running very busy
-journals centralization servers aggregating several thousands of log entries per second, the number of files can grow
-to several dozens of thousands quickly. In such setups, the plugin should ideally skip processing journal files
-entirely, relying solely on the estimations of the sequence of files each file is part of. However, this has not been
-implemented yet. To improve the query performance in such setups, the user has to query smaller timeframes.
-
-Another optimization taking place in huge journal centralization points, is the initial scan of the database. The plugin
-needs to know the list of all journal files available, including the details of the first and the last message in each
-of them. When there are several thousands of files in a directory (like it usually happens in `/var/log/journal/remote`),
-directory listing and examination of each file can take a considerable amount of time (even `ls -l` takes minutes).
-To work around this problem, the plugin uses `inotify` to receive file updates immediately and scans the library from
-the newest to the oldest file, allowing the user interface to work immediately after startup, for the most recent
-timeframes.
-
-### Best practices for better performance
-
-systemd-journal has been designed **first to be reliable** and then to be fast. It includes several mechanisms to ensure
-minimal data loss under all conditions (e.g. disk corruption, tampering, forward secure sealing) and despite the fact
-that it utilizes several techniques to require minimal disk footprint (like deduplication of log entries, linking of
-values and fields, compression) the disk footprint of journal files remains significantly higher compared to other log
-management solutions.
-
-The higher disk footprint results in higher disk I/O during querying, since a lot more data have to read from disk to
-evaluate a query. Query performance at scale can greatly benefit by utilizing a compressed filesystem (ext4, btrfs, zfs)
-to store systemd-journal files.
-
-systemd-journal files are cached by the operating system. There is no database server to serve queries. Each file is
-opened and the query runs by directly accessing the data in it.
-
-Therefore systemd-journal relies on the caching layer of the operating system to optimize query performance. The more
-RAM the system has, although it will not be reported as `used` (it will be reported as `cache`), the faster the queries
-will get. The first time a timeframe is accessed the query performance will be slower, but further queries on the same
-timeframe will be significantly faster since journal data are now cached in memory.
-
-So, on busy logs centralization systems, queries performance can be improved significantly by using a compressed
-filesystem for storing the journal files, and higher amounts of RAM.
-
-## Configuration and maintenance
-
-This Netdata plugin does not require any configuration or maintenance.
-
-## FAQ
-
-### Can I use this plugin on journal centralization servers?
-
-Yes. You can centralize your logs using `systemd-journal-remote`, and then install Netdata
-on this logs centralization server to explore the logs of all your infrastructure.
-
-This plugin will automatically provide multi-node views of your logs and also give you the ability to combine the logs
-of multiple servers, as you see fit.
-
-Check [configuring a logs centralization server](#how-do-i-configure-a-journal-centralization-server).
-
-### Can I use this plugin from a parent Netdata?
-
-Yes. When your nodes are connected to a Netdata parent, all their functions are available
-via the parent's UI. So, from the parent UI, you can access the functions of all your nodes.
-
-Keep in mind that to protect your privacy, in order to access Netdata functions, you need a
-free Netdata Cloud account.
-
-### Is any of my data exposed to Netdata Cloud from this plugin?
-
-No. When you access the agent directly, none of your data passes through Netdata Cloud.
-You need a free Netdata Cloud account only to verify your identity and enable the use of
-Netdata Functions. Once this is done, all the data flow directly from your Netdata agent
-to your web browser.
-
-Also check [this discussion](https://github.com/netdata/netdata/discussions/16136).
-
-When you access Netdata via `https://app.netdata.cloud`, your data travel via Netdata Cloud,
-but they are not stored in Netdata Cloud. This is to allow you access your Netdata agents from
-anywhere. All communication from/to Netdata Cloud is encrypted.
-
-### What are `volatile` and `persistent` journals?
-
-`systemd` `journald` allows creating both `volatile` journals in a `tmpfs` ram drive,
-and `persistent` journals stored on disk.
-
-`volatile` journals are particularly useful when the system monitored is sensitive to
-disk I/O, or does not have any writable disks at all.
-
-For more information check `man systemd-journald`.
-
-### I centralize my logs with Loki. Why to use Netdata for my journals?
-
-`systemd` journals have almost infinite cardinality at their labels and all of them are indexed,
-even if every single message has unique fields and values.
-
-When you send `systemd` journal logs to Loki, even if you use the `relabel_rules` argument to
-`loki.source.journal` with a JSON format, you need to specify which of the fields from journald
-you want inherited by Loki. This means you need to know the most important fields beforehand.
-At the same time you loose all the flexibility `systemd` journal provides:
-**indexing on all fields and all their values**.
-
-Loki generally assumes that all logs are like a table. All entries in a stream share the same
-fields. But journald does exactly the opposite. Each log entry is unique and may have its own unique fields.
-
-So, Loki and `systemd-journal` are good for different use cases.
-
-`systemd-journal` already runs in your systems. You use it today. It is there inside all your systems
-collecting the system and applications logs. And for its use case, it has advantages over other
-centralization solutions. So, why not use it?
-
-### Is it worth to build a `systemd` logs centralization server?
-
-Yes. It is simple, fast and the software to do it is already in your systems.
-
-For application and system logs, `systemd` journal is ideal and the visibility you can get
-by centralizing your system logs and the use of this Netdata plugin, is unparalleled.
-
-### How do I configure a journal centralization server?
-
-A short summary to get journal server running can be found below.
-There are two strategies you can apply, when it comes down to a centralized server for `systemd` journal logs.
-
-1. _Active sources_, where the centralized server fetches the logs from each individual server
-2. _Passive sources_, where the centralized server accepts a log stream from an individual server.
-
-For more options and reference to documentation, check `man systemd-journal-remote` and `man systemd-journal-upload`.
-
-#### _passive_ journal centralization without encryption
-
-If you want to setup your own passive journal centralization setup without encryption, [check out guide on it](https://github.com/netdata/netdata/blob/master/collectors/systemd-journal.plugin/passive_journal_centralization_guide_no_encryption.md).
-
-#### _passive_ journal centralization with encryption using self-signed certificates
-
-If you want to setup your own passive journal centralization setup using self-signed certificates for encryption, [check out guide on it](https://github.com/netdata/netdata/blob/master/collectors/systemd-journal.plugin/passive_journal_centralization_guide_self_signed_certs.md).
-
-#### Limitations when using a logs centralization server
-
-As of this writing `namespaces` support by `systemd` is limited:
-
-- Docker containers cannot log to namespaces. Check [this issue](https://github.com/moby/moby/issues/41879).
-- `systemd-journal-upload` automatically uploads `system` and `user` journals, but not `namespaces` journals. For this
- you need to spawn a `systemd-journal-upload` per namespace.
diff --git a/collectors/systemd-journal.plugin/systemd-journal-annotations.c b/collectors/systemd-journal.plugin/systemd-journal-annotations.c
deleted file mode 100644
index b12356110..000000000
--- a/collectors/systemd-journal.plugin/systemd-journal-annotations.c
+++ /dev/null
@@ -1,719 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-#include "systemd-internals.h"
-
-const char *errno_map[] = {
- [1] = "1 (EPERM)", // "Operation not permitted",
- [2] = "2 (ENOENT)", // "No such file or directory",
- [3] = "3 (ESRCH)", // "No such process",
- [4] = "4 (EINTR)", // "Interrupted system call",
- [5] = "5 (EIO)", // "Input/output error",
- [6] = "6 (ENXIO)", // "No such device or address",
- [7] = "7 (E2BIG)", // "Argument list too long",
- [8] = "8 (ENOEXEC)", // "Exec format error",
- [9] = "9 (EBADF)", // "Bad file descriptor",
- [10] = "10 (ECHILD)", // "No child processes",
- [11] = "11 (EAGAIN)", // "Resource temporarily unavailable",
- [12] = "12 (ENOMEM)", // "Cannot allocate memory",
- [13] = "13 (EACCES)", // "Permission denied",
- [14] = "14 (EFAULT)", // "Bad address",
- [15] = "15 (ENOTBLK)", // "Block device required",
- [16] = "16 (EBUSY)", // "Device or resource busy",
- [17] = "17 (EEXIST)", // "File exists",
- [18] = "18 (EXDEV)", // "Invalid cross-device link",
- [19] = "19 (ENODEV)", // "No such device",
- [20] = "20 (ENOTDIR)", // "Not a directory",
- [21] = "21 (EISDIR)", // "Is a directory",
- [22] = "22 (EINVAL)", // "Invalid argument",
- [23] = "23 (ENFILE)", // "Too many open files in system",
- [24] = "24 (EMFILE)", // "Too many open files",
- [25] = "25 (ENOTTY)", // "Inappropriate ioctl for device",
- [26] = "26 (ETXTBSY)", // "Text file busy",
- [27] = "27 (EFBIG)", // "File too large",
- [28] = "28 (ENOSPC)", // "No space left on device",
- [29] = "29 (ESPIPE)", // "Illegal seek",
- [30] = "30 (EROFS)", // "Read-only file system",
- [31] = "31 (EMLINK)", // "Too many links",
- [32] = "32 (EPIPE)", // "Broken pipe",
- [33] = "33 (EDOM)", // "Numerical argument out of domain",
- [34] = "34 (ERANGE)", // "Numerical result out of range",
- [35] = "35 (EDEADLK)", // "Resource deadlock avoided",
- [36] = "36 (ENAMETOOLONG)", // "File name too long",
- [37] = "37 (ENOLCK)", // "No locks available",
- [38] = "38 (ENOSYS)", // "Function not implemented",
- [39] = "39 (ENOTEMPTY)", // "Directory not empty",
- [40] = "40 (ELOOP)", // "Too many levels of symbolic links",
- [42] = "42 (ENOMSG)", // "No message of desired type",
- [43] = "43 (EIDRM)", // "Identifier removed",
- [44] = "44 (ECHRNG)", // "Channel number out of range",
- [45] = "45 (EL2NSYNC)", // "Level 2 not synchronized",
- [46] = "46 (EL3HLT)", // "Level 3 halted",
- [47] = "47 (EL3RST)", // "Level 3 reset",
- [48] = "48 (ELNRNG)", // "Link number out of range",
- [49] = "49 (EUNATCH)", // "Protocol driver not attached",
- [50] = "50 (ENOCSI)", // "No CSI structure available",
- [51] = "51 (EL2HLT)", // "Level 2 halted",
- [52] = "52 (EBADE)", // "Invalid exchange",
- [53] = "53 (EBADR)", // "Invalid request descriptor",
- [54] = "54 (EXFULL)", // "Exchange full",
- [55] = "55 (ENOANO)", // "No anode",
- [56] = "56 (EBADRQC)", // "Invalid request code",
- [57] = "57 (EBADSLT)", // "Invalid slot",
- [59] = "59 (EBFONT)", // "Bad font file format",
- [60] = "60 (ENOSTR)", // "Device not a stream",
- [61] = "61 (ENODATA)", // "No data available",
- [62] = "62 (ETIME)", // "Timer expired",
- [63] = "63 (ENOSR)", // "Out of streams resources",
- [64] = "64 (ENONET)", // "Machine is not on the network",
- [65] = "65 (ENOPKG)", // "Package not installed",
- [66] = "66 (EREMOTE)", // "Object is remote",
- [67] = "67 (ENOLINK)", // "Link has been severed",
- [68] = "68 (EADV)", // "Advertise error",
- [69] = "69 (ESRMNT)", // "Srmount error",
- [70] = "70 (ECOMM)", // "Communication error on send",
- [71] = "71 (EPROTO)", // "Protocol error",
- [72] = "72 (EMULTIHOP)", // "Multihop attempted",
- [73] = "73 (EDOTDOT)", // "RFS specific error",
- [74] = "74 (EBADMSG)", // "Bad message",
- [75] = "75 (EOVERFLOW)", // "Value too large for defined data type",
- [76] = "76 (ENOTUNIQ)", // "Name not unique on network",
- [77] = "77 (EBADFD)", // "File descriptor in bad state",
- [78] = "78 (EREMCHG)", // "Remote address changed",
- [79] = "79 (ELIBACC)", // "Can not access a needed shared library",
- [80] = "80 (ELIBBAD)", // "Accessing a corrupted shared library",
- [81] = "81 (ELIBSCN)", // ".lib section in a.out corrupted",
- [82] = "82 (ELIBMAX)", // "Attempting to link in too many shared libraries",
- [83] = "83 (ELIBEXEC)", // "Cannot exec a shared library directly",
- [84] = "84 (EILSEQ)", // "Invalid or incomplete multibyte or wide character",
- [85] = "85 (ERESTART)", // "Interrupted system call should be restarted",
- [86] = "86 (ESTRPIPE)", // "Streams pipe error",
- [87] = "87 (EUSERS)", // "Too many users",
- [88] = "88 (ENOTSOCK)", // "Socket operation on non-socket",
- [89] = "89 (EDESTADDRREQ)", // "Destination address required",
- [90] = "90 (EMSGSIZE)", // "Message too long",
- [91] = "91 (EPROTOTYPE)", // "Protocol wrong type for socket",
- [92] = "92 (ENOPROTOOPT)", // "Protocol not available",
- [93] = "93 (EPROTONOSUPPORT)", // "Protocol not supported",
- [94] = "94 (ESOCKTNOSUPPORT)", // "Socket type not supported",
- [95] = "95 (ENOTSUP)", // "Operation not supported",
- [96] = "96 (EPFNOSUPPORT)", // "Protocol family not supported",
- [97] = "97 (EAFNOSUPPORT)", // "Address family not supported by protocol",
- [98] = "98 (EADDRINUSE)", // "Address already in use",
- [99] = "99 (EADDRNOTAVAIL)", // "Cannot assign requested address",
- [100] = "100 (ENETDOWN)", // "Network is down",
- [101] = "101 (ENETUNREACH)", // "Network is unreachable",
- [102] = "102 (ENETRESET)", // "Network dropped connection on reset",
- [103] = "103 (ECONNABORTED)", // "Software caused connection abort",
- [104] = "104 (ECONNRESET)", // "Connection reset by peer",
- [105] = "105 (ENOBUFS)", // "No buffer space available",
- [106] = "106 (EISCONN)", // "Transport endpoint is already connected",
- [107] = "107 (ENOTCONN)", // "Transport endpoint is not connected",
- [108] = "108 (ESHUTDOWN)", // "Cannot send after transport endpoint shutdown",
- [109] = "109 (ETOOMANYREFS)", // "Too many references: cannot splice",
- [110] = "110 (ETIMEDOUT)", // "Connection timed out",
- [111] = "111 (ECONNREFUSED)", // "Connection refused",
- [112] = "112 (EHOSTDOWN)", // "Host is down",
- [113] = "113 (EHOSTUNREACH)", // "No route to host",
- [114] = "114 (EALREADY)", // "Operation already in progress",
- [115] = "115 (EINPROGRESS)", // "Operation now in progress",
- [116] = "116 (ESTALE)", // "Stale file handle",
- [117] = "117 (EUCLEAN)", // "Structure needs cleaning",
- [118] = "118 (ENOTNAM)", // "Not a XENIX named type file",
- [119] = "119 (ENAVAIL)", // "No XENIX semaphores available",
- [120] = "120 (EISNAM)", // "Is a named type file",
- [121] = "121 (EREMOTEIO)", // "Remote I/O error",
- [122] = "122 (EDQUOT)", // "Disk quota exceeded",
- [123] = "123 (ENOMEDIUM)", // "No medium found",
- [124] = "124 (EMEDIUMTYPE)", // "Wrong medium type",
- [125] = "125 (ECANCELED)", // "Operation canceled",
- [126] = "126 (ENOKEY)", // "Required key not available",
- [127] = "127 (EKEYEXPIRED)", // "Key has expired",
- [128] = "128 (EKEYREVOKED)", // "Key has been revoked",
- [129] = "129 (EKEYREJECTED)", // "Key was rejected by service",
- [130] = "130 (EOWNERDEAD)", // "Owner died",
- [131] = "131 (ENOTRECOVERABLE)", // "State not recoverable",
- [132] = "132 (ERFKILL)", // "Operation not possible due to RF-kill",
- [133] = "133 (EHWPOISON)", // "Memory page has hardware error",
-};
-
-const char *linux_capabilities[] = {
- [CAP_CHOWN] = "CHOWN",
- [CAP_DAC_OVERRIDE] = "DAC_OVERRIDE",
- [CAP_DAC_READ_SEARCH] = "DAC_READ_SEARCH",
- [CAP_FOWNER] = "FOWNER",
- [CAP_FSETID] = "FSETID",
- [CAP_KILL] = "KILL",
- [CAP_SETGID] = "SETGID",
- [CAP_SETUID] = "SETUID",
- [CAP_SETPCAP] = "SETPCAP",
- [CAP_LINUX_IMMUTABLE] = "LINUX_IMMUTABLE",
- [CAP_NET_BIND_SERVICE] = "NET_BIND_SERVICE",
- [CAP_NET_BROADCAST] = "NET_BROADCAST",
- [CAP_NET_ADMIN] = "NET_ADMIN",
- [CAP_NET_RAW] = "NET_RAW",
- [CAP_IPC_LOCK] = "IPC_LOCK",
- [CAP_IPC_OWNER] = "IPC_OWNER",
- [CAP_SYS_MODULE] = "SYS_MODULE",
- [CAP_SYS_RAWIO] = "SYS_RAWIO",
- [CAP_SYS_CHROOT] = "SYS_CHROOT",
- [CAP_SYS_PTRACE] = "SYS_PTRACE",
- [CAP_SYS_PACCT] = "SYS_PACCT",
- [CAP_SYS_ADMIN] = "SYS_ADMIN",
- [CAP_SYS_BOOT] = "SYS_BOOT",
- [CAP_SYS_NICE] = "SYS_NICE",
- [CAP_SYS_RESOURCE] = "SYS_RESOURCE",
- [CAP_SYS_TIME] = "SYS_TIME",
- [CAP_SYS_TTY_CONFIG] = "SYS_TTY_CONFIG",
- [CAP_MKNOD] = "MKNOD",
- [CAP_LEASE] = "LEASE",
- [CAP_AUDIT_WRITE] = "AUDIT_WRITE",
- [CAP_AUDIT_CONTROL] = "AUDIT_CONTROL",
- [CAP_SETFCAP] = "SETFCAP",
- [CAP_MAC_OVERRIDE] = "MAC_OVERRIDE",
- [CAP_MAC_ADMIN] = "MAC_ADMIN",
- [CAP_SYSLOG] = "SYSLOG",
- [CAP_WAKE_ALARM] = "WAKE_ALARM",
- [CAP_BLOCK_SUSPEND] = "BLOCK_SUSPEND",
- [37 /*CAP_AUDIT_READ*/] = "AUDIT_READ",
- [38 /*CAP_PERFMON*/] = "PERFMON",
- [39 /*CAP_BPF*/] = "BPF",
- [40 /* CAP_CHECKPOINT_RESTORE */] = "CHECKPOINT_RESTORE",
-};
-
-static const char *syslog_facility_to_name(int facility) {
- switch (facility) {
- case LOG_FAC(LOG_KERN): return "kern";
- case LOG_FAC(LOG_USER): return "user";
- case LOG_FAC(LOG_MAIL): return "mail";
- case LOG_FAC(LOG_DAEMON): return "daemon";
- case LOG_FAC(LOG_AUTH): return "auth";
- case LOG_FAC(LOG_SYSLOG): return "syslog";
- case LOG_FAC(LOG_LPR): return "lpr";
- case LOG_FAC(LOG_NEWS): return "news";
- case LOG_FAC(LOG_UUCP): return "uucp";
- case LOG_FAC(LOG_CRON): return "cron";
- case LOG_FAC(LOG_AUTHPRIV): return "authpriv";
- case LOG_FAC(LOG_FTP): return "ftp";
- case LOG_FAC(LOG_LOCAL0): return "local0";
- case LOG_FAC(LOG_LOCAL1): return "local1";
- case LOG_FAC(LOG_LOCAL2): return "local2";
- case LOG_FAC(LOG_LOCAL3): return "local3";
- case LOG_FAC(LOG_LOCAL4): return "local4";
- case LOG_FAC(LOG_LOCAL5): return "local5";
- case LOG_FAC(LOG_LOCAL6): return "local6";
- case LOG_FAC(LOG_LOCAL7): return "local7";
- default: return NULL;
- }
-}
-
-static const char *syslog_priority_to_name(int priority) {
- switch (priority) {
- case LOG_ALERT: return "alert";
- case LOG_CRIT: return "critical";
- case LOG_DEBUG: return "debug";
- case LOG_EMERG: return "panic";
- case LOG_ERR: return "error";
- case LOG_INFO: return "info";
- case LOG_NOTICE: return "notice";
- case LOG_WARNING: return "warning";
- default: return NULL;
- }
-}
-
-FACET_ROW_SEVERITY syslog_priority_to_facet_severity(FACETS *facets __maybe_unused, FACET_ROW *row, void *data __maybe_unused) {
- // same to
- // https://github.com/systemd/systemd/blob/aab9e4b2b86905a15944a1ac81e471b5b7075932/src/basic/terminal-util.c#L1501
- // function get_log_colors()
-
- FACET_ROW_KEY_VALUE *priority_rkv = dictionary_get(row->dict, "PRIORITY");
- if(!priority_rkv || priority_rkv->empty)
- return FACET_ROW_SEVERITY_NORMAL;
-
- int priority = str2i(buffer_tostring(priority_rkv->wb));
-
- if(priority <= LOG_ERR)
- return FACET_ROW_SEVERITY_CRITICAL;
-
- else if (priority <= LOG_WARNING)
- return FACET_ROW_SEVERITY_WARNING;
-
- else if(priority <= LOG_NOTICE)
- return FACET_ROW_SEVERITY_NOTICE;
-
- else if(priority >= LOG_DEBUG)
- return FACET_ROW_SEVERITY_DEBUG;
-
- return FACET_ROW_SEVERITY_NORMAL;
-}
-
-static char *uid_to_username(uid_t uid, char *buffer, size_t buffer_size) {
- static __thread char tmp[1024 + 1];
- struct passwd pw, *result = NULL;
-
- if (getpwuid_r(uid, &pw, tmp, sizeof(tmp), &result) != 0 || !result || !pw.pw_name || !(*pw.pw_name))
- snprintfz(buffer, buffer_size - 1, "%u", uid);
- else
- snprintfz(buffer, buffer_size - 1, "%u (%s)", uid, pw.pw_name);
-
- return buffer;
-}
-
-static char *gid_to_groupname(gid_t gid, char* buffer, size_t buffer_size) {
- static __thread char tmp[1024];
- struct group grp, *result = NULL;
-
- if (getgrgid_r(gid, &grp, tmp, sizeof(tmp), &result) != 0 || !result || !grp.gr_name || !(*grp.gr_name))
- snprintfz(buffer, buffer_size - 1, "%u", gid);
- else
- snprintfz(buffer, buffer_size - 1, "%u (%s)", gid, grp.gr_name);
-
- return buffer;
-}
-
-void netdata_systemd_journal_transform_syslog_facility(FACETS *facets __maybe_unused, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope __maybe_unused, void *data __maybe_unused) {
- const char *v = buffer_tostring(wb);
- if(*v && isdigit(*v)) {
- int facility = str2i(buffer_tostring(wb));
- const char *name = syslog_facility_to_name(facility);
- if (name) {
- buffer_flush(wb);
- buffer_strcat(wb, name);
- }
- }
-}
-
-void netdata_systemd_journal_transform_priority(FACETS *facets __maybe_unused, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope __maybe_unused, void *data __maybe_unused) {
- if(scope == FACETS_TRANSFORM_FACET_SORT)
- return;
-
- const char *v = buffer_tostring(wb);
- if(*v && isdigit(*v)) {
- int priority = str2i(buffer_tostring(wb));
- const char *name = syslog_priority_to_name(priority);
- if (name) {
- buffer_flush(wb);
- buffer_strcat(wb, name);
- }
- }
-}
-
-void netdata_systemd_journal_transform_errno(FACETS *facets __maybe_unused, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope __maybe_unused, void *data __maybe_unused) {
- if(scope == FACETS_TRANSFORM_FACET_SORT)
- return;
-
- const char *v = buffer_tostring(wb);
- if(*v && isdigit(*v)) {
- unsigned err_no = str2u(buffer_tostring(wb));
- if(err_no > 0 && err_no < sizeof(errno_map) / sizeof(*errno_map)) {
- const char *name = errno_map[err_no];
- if(name) {
- buffer_flush(wb);
- buffer_strcat(wb, name);
- }
- }
- }
-}
-
-// ----------------------------------------------------------------------------
-// UID and GID transformation
-
-#define UID_GID_HASHTABLE_SIZE 10000
-
-struct word_t2str_hashtable_entry {
- struct word_t2str_hashtable_entry *next;
- Word_t hash;
- size_t len;
- char str[];
-};
-
-struct word_t2str_hashtable {
- SPINLOCK spinlock;
- size_t size;
- struct word_t2str_hashtable_entry *hashtable[UID_GID_HASHTABLE_SIZE];
-};
-
-struct word_t2str_hashtable uid_hashtable = {
- .size = UID_GID_HASHTABLE_SIZE,
-};
-
-struct word_t2str_hashtable gid_hashtable = {
- .size = UID_GID_HASHTABLE_SIZE,
-};
-
-struct word_t2str_hashtable_entry **word_t2str_hashtable_slot(struct word_t2str_hashtable *ht, Word_t hash) {
- size_t slot = hash % ht->size;
- struct word_t2str_hashtable_entry **e = &ht->hashtable[slot];
-
- while(*e && (*e)->hash != hash)
- e = &((*e)->next);
-
- return e;
-}
-
-const char *uid_to_username_cached(uid_t uid, size_t *length) {
- spinlock_lock(&uid_hashtable.spinlock);
-
- struct word_t2str_hashtable_entry **e = word_t2str_hashtable_slot(&uid_hashtable, uid);
- if(!(*e)) {
- static __thread char buf[1024];
- const char *name = uid_to_username(uid, buf, sizeof(buf));
- size_t size = strlen(name) + 1;
-
- *e = callocz(1, sizeof(struct word_t2str_hashtable_entry) + size);
- (*e)->len = size - 1;
- (*e)->hash = uid;
- memcpy((*e)->str, name, size);
- }
-
- spinlock_unlock(&uid_hashtable.spinlock);
-
- *length = (*e)->len;
- return (*e)->str;
-}
-
-const char *gid_to_groupname_cached(gid_t gid, size_t *length) {
- spinlock_lock(&gid_hashtable.spinlock);
-
- struct word_t2str_hashtable_entry **e = word_t2str_hashtable_slot(&gid_hashtable, gid);
- if(!(*e)) {
- static __thread char buf[1024];
- const char *name = gid_to_groupname(gid, buf, sizeof(buf));
- size_t size = strlen(name) + 1;
-
- *e = callocz(1, sizeof(struct word_t2str_hashtable_entry) + size);
- (*e)->len = size - 1;
- (*e)->hash = gid;
- memcpy((*e)->str, name, size);
- }
-
- spinlock_unlock(&gid_hashtable.spinlock);
-
- *length = (*e)->len;
- return (*e)->str;
-}
-
-DICTIONARY *boot_ids_to_first_ut = NULL;
-
-void netdata_systemd_journal_transform_boot_id(FACETS *facets __maybe_unused, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope __maybe_unused, void *data __maybe_unused) {
- const char *boot_id = buffer_tostring(wb);
- if(*boot_id && isxdigit(*boot_id)) {
- usec_t ut = UINT64_MAX;
- usec_t *p_ut = dictionary_get(boot_ids_to_first_ut, boot_id);
- if(!p_ut) {
-#ifndef HAVE_SD_JOURNAL_RESTART_FIELDS
- struct journal_file *jf;
- dfe_start_read(journal_files_registry, jf) {
- const char *files[2] = {
- [0] = jf_dfe.name,
- [1] = NULL,
- };
-
- sd_journal *j = NULL;
- int r = sd_journal_open_files(&j, files, ND_SD_JOURNAL_OPEN_FLAGS);
- if(r < 0 || !j) {
- internal_error(true, "JOURNAL: while looking for the first timestamp of boot_id '%s', "
- "sd_journal_open_files('%s') returned %d",
- boot_id, jf_dfe.name, r);
- continue;
- }
-
- ut = journal_file_update_annotation_boot_id(j, jf, boot_id);
- sd_journal_close(j);
- }
- dfe_done(jf);
-#endif
- }
- else
- ut = *p_ut;
-
- if(ut && ut != UINT64_MAX) {
- char buffer[RFC3339_MAX_LENGTH];
- rfc3339_datetime_ut(buffer, sizeof(buffer), ut, 0, true);
-
- switch(scope) {
- default:
- case FACETS_TRANSFORM_DATA:
- case FACETS_TRANSFORM_VALUE:
- buffer_sprintf(wb, " (%s) ", buffer);
- break;
-
- case FACETS_TRANSFORM_FACET:
- case FACETS_TRANSFORM_FACET_SORT:
- case FACETS_TRANSFORM_HISTOGRAM:
- buffer_flush(wb);
- buffer_sprintf(wb, "%s", buffer);
- break;
- }
- }
- }
-}
-
-void netdata_systemd_journal_transform_uid(FACETS *facets __maybe_unused, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope __maybe_unused, void *data __maybe_unused) {
- if(scope == FACETS_TRANSFORM_FACET_SORT)
- return;
-
- const char *v = buffer_tostring(wb);
- if(*v && isdigit(*v)) {
- uid_t uid = str2i(buffer_tostring(wb));
- size_t len;
- const char *name = uid_to_username_cached(uid, &len);
- buffer_contents_replace(wb, name, len);
- }
-}
-
-void netdata_systemd_journal_transform_gid(FACETS *facets __maybe_unused, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope __maybe_unused, void *data __maybe_unused) {
- if(scope == FACETS_TRANSFORM_FACET_SORT)
- return;
-
- const char *v = buffer_tostring(wb);
- if(*v && isdigit(*v)) {
- gid_t gid = str2i(buffer_tostring(wb));
- size_t len;
- const char *name = gid_to_groupname_cached(gid, &len);
- buffer_contents_replace(wb, name, len);
- }
-}
-
-void netdata_systemd_journal_transform_cap_effective(FACETS *facets __maybe_unused, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope __maybe_unused, void *data __maybe_unused) {
- if(scope == FACETS_TRANSFORM_FACET_SORT)
- return;
-
- const char *v = buffer_tostring(wb);
- if(*v && isdigit(*v)) {
- uint64_t cap = strtoul(buffer_tostring(wb), NULL, 16);
- if(cap) {
- buffer_fast_strcat(wb, " (", 2);
- for (size_t i = 0, added = 0; i < sizeof(linux_capabilities) / sizeof(linux_capabilities[0]); i++) {
- if (linux_capabilities[i] && (cap & (1ULL << i))) {
-
- if (added)
- buffer_fast_strcat(wb, " | ", 3);
-
- buffer_strcat(wb, linux_capabilities[i]);
- added++;
- }
- }
- buffer_fast_strcat(wb, ")", 1);
- }
- }
-}
-
-void netdata_systemd_journal_transform_timestamp_usec(FACETS *facets __maybe_unused, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope __maybe_unused, void *data __maybe_unused) {
- if(scope == FACETS_TRANSFORM_FACET_SORT)
- return;
-
- const char *v = buffer_tostring(wb);
- if(*v && isdigit(*v)) {
- uint64_t ut = str2ull(buffer_tostring(wb), NULL);
- if(ut) {
- char buffer[RFC3339_MAX_LENGTH];
- rfc3339_datetime_ut(buffer, sizeof(buffer), ut, 6, true);
- buffer_sprintf(wb, " (%s)", buffer);
- }
- }
-}
-
-// ----------------------------------------------------------------------------
-
-void netdata_systemd_journal_dynamic_row_id(FACETS *facets __maybe_unused, BUFFER *json_array, FACET_ROW_KEY_VALUE *rkv, FACET_ROW *row, void *data __maybe_unused) {
- FACET_ROW_KEY_VALUE *pid_rkv = dictionary_get(row->dict, "_PID");
- const char *pid = pid_rkv ? buffer_tostring(pid_rkv->wb) : FACET_VALUE_UNSET;
-
- const char *identifier = NULL;
- FACET_ROW_KEY_VALUE *container_name_rkv = dictionary_get(row->dict, "CONTAINER_NAME");
- if(container_name_rkv && !container_name_rkv->empty)
- identifier = buffer_tostring(container_name_rkv->wb);
-
- if(!identifier) {
- FACET_ROW_KEY_VALUE *syslog_identifier_rkv = dictionary_get(row->dict, "SYSLOG_IDENTIFIER");
- if(syslog_identifier_rkv && !syslog_identifier_rkv->empty)
- identifier = buffer_tostring(syslog_identifier_rkv->wb);
-
- if(!identifier) {
- FACET_ROW_KEY_VALUE *comm_rkv = dictionary_get(row->dict, "_COMM");
- if(comm_rkv && !comm_rkv->empty)
- identifier = buffer_tostring(comm_rkv->wb);
- }
- }
-
- buffer_flush(rkv->wb);
-
- if(!identifier || !*identifier)
- buffer_strcat(rkv->wb, FACET_VALUE_UNSET);
- else if(!pid || !*pid)
- buffer_sprintf(rkv->wb, "%s", identifier);
- else
- buffer_sprintf(rkv->wb, "%s[%s]", identifier, pid);
-
- buffer_json_add_array_item_string(json_array, buffer_tostring(rkv->wb));
-}
-
-
-// ----------------------------------------------------------------------------
-
-struct message_id_info {
- const char *msg;
-};
-
-static DICTIONARY *known_journal_messages_ids = NULL;
-
-void netdata_systemd_journal_message_ids_init(void) {
- known_journal_messages_ids = dictionary_create(DICT_OPTION_DONT_OVERWRITE_VALUE);
-
- struct message_id_info i = { 0 };
- i.msg = "Journal start"; dictionary_set(known_journal_messages_ids, "f77379a8490b408bbe5f6940505a777b", &i, sizeof(i));
- i.msg = "Journal stop"; dictionary_set(known_journal_messages_ids, "d93fb3c9c24d451a97cea615ce59c00b", &i, sizeof(i));
- i.msg = "Journal dropped"; dictionary_set(known_journal_messages_ids, "a596d6fe7bfa4994828e72309e95d61e", &i, sizeof(i));
- i.msg = "Journal missed"; dictionary_set(known_journal_messages_ids, "e9bf28e6e834481bb6f48f548ad13606", &i, sizeof(i));
- i.msg = "Journal usage"; dictionary_set(known_journal_messages_ids, "ec387f577b844b8fa948f33cad9a75e6", &i, sizeof(i));
- i.msg = "Coredump"; dictionary_set(known_journal_messages_ids, "fc2e22bc6ee647b6b90729ab34a250b1", &i, sizeof(i));
- i.msg = "Truncated core"; dictionary_set(known_journal_messages_ids, "5aadd8e954dc4b1a8c954d63fd9e1137", &i, sizeof(i));
- i.msg = "Backtrace"; dictionary_set(known_journal_messages_ids, "1f4e0a44a88649939aaea34fc6da8c95", &i, sizeof(i));
- i.msg = "Session start"; dictionary_set(known_journal_messages_ids, "8d45620c1a4348dbb17410da57c60c66", &i, sizeof(i));
- i.msg = "Session stop"; dictionary_set(known_journal_messages_ids, "3354939424b4456d9802ca8333ed424a", &i, sizeof(i));
- i.msg = "Seat start"; dictionary_set(known_journal_messages_ids, "fcbefc5da23d428093f97c82a9290f7b", &i, sizeof(i));
- i.msg = "Seat stop"; dictionary_set(known_journal_messages_ids, "e7852bfe46784ed0accde04bc864c2d5", &i, sizeof(i));
- i.msg = "Machine start"; dictionary_set(known_journal_messages_ids, "24d8d4452573402496068381a6312df2", &i, sizeof(i));
- i.msg = "Machine stop"; dictionary_set(known_journal_messages_ids, "58432bd3bace477cb514b56381b8a758", &i, sizeof(i));
- i.msg = "Time change"; dictionary_set(known_journal_messages_ids, "c7a787079b354eaaa9e77b371893cd27", &i, sizeof(i));
- i.msg = "Timezone change"; dictionary_set(known_journal_messages_ids, "45f82f4aef7a4bbf942ce861d1f20990", &i, sizeof(i));
- i.msg = "Tainted"; dictionary_set(known_journal_messages_ids, "50876a9db00f4c40bde1a2ad381c3a1b", &i, sizeof(i));
- i.msg = "Startup finished"; dictionary_set(known_journal_messages_ids, "b07a249cd024414a82dd00cd181378ff", &i, sizeof(i));
- i.msg = "User startup finished"; dictionary_set(known_journal_messages_ids, "eed00a68ffd84e31882105fd973abdd1", &i, sizeof(i));
- i.msg = "Sleep start"; dictionary_set(known_journal_messages_ids, "6bbd95ee977941e497c48be27c254128", &i, sizeof(i));
- i.msg = "Sleep stop"; dictionary_set(known_journal_messages_ids, "8811e6df2a8e40f58a94cea26f8ebf14", &i, sizeof(i));
- i.msg = "Shutdown"; dictionary_set(known_journal_messages_ids, "98268866d1d54a499c4e98921d93bc40", &i, sizeof(i));
- i.msg = "Factory reset"; dictionary_set(known_journal_messages_ids, "c14aaf76ec284a5fa1f105f88dfb061c", &i, sizeof(i));
- i.msg = "Crash exit"; dictionary_set(known_journal_messages_ids, "d9ec5e95e4b646aaaea2fd05214edbda", &i, sizeof(i));
- i.msg = "Crash failed"; dictionary_set(known_journal_messages_ids, "3ed0163e868a4417ab8b9e210407a96c", &i, sizeof(i));
- i.msg = "Crash freeze"; dictionary_set(known_journal_messages_ids, "645c735537634ae0a32b15a7c6cba7d4", &i, sizeof(i));
- i.msg = "Crash no coredump"; dictionary_set(known_journal_messages_ids, "5addb3a06a734d3396b794bf98fb2d01", &i, sizeof(i));
- i.msg = "Crash no fork"; dictionary_set(known_journal_messages_ids, "5c9e98de4ab94c6a9d04d0ad793bd903", &i, sizeof(i));
- i.msg = "Crash unknown signal"; dictionary_set(known_journal_messages_ids, "5e6f1f5e4db64a0eaee3368249d20b94", &i, sizeof(i));
- i.msg = "Crash systemd signal"; dictionary_set(known_journal_messages_ids, "83f84b35ee264f74a3896a9717af34cb", &i, sizeof(i));
- i.msg = "Crash process signal"; dictionary_set(known_journal_messages_ids, "3a73a98baf5b4b199929e3226c0be783", &i, sizeof(i));
- i.msg = "Crash waitpid failed"; dictionary_set(known_journal_messages_ids, "2ed18d4f78ca47f0a9bc25271c26adb4", &i, sizeof(i));
- i.msg = "Crash coredump failed"; dictionary_set(known_journal_messages_ids, "56b1cd96f24246c5b607666fda952356", &i, sizeof(i));
- i.msg = "Crash coredump pid"; dictionary_set(known_journal_messages_ids, "4ac7566d4d7548f4981f629a28f0f829", &i, sizeof(i));
- i.msg = "Crash shell fork failed"; dictionary_set(known_journal_messages_ids, "38e8b1e039ad469291b18b44c553a5b7", &i, sizeof(i));
- i.msg = "Crash execle failed"; dictionary_set(known_journal_messages_ids, "872729b47dbe473eb768ccecd477beda", &i, sizeof(i));
- i.msg = "Selinux failed"; dictionary_set(known_journal_messages_ids, "658a67adc1c940b3b3316e7e8628834a", &i, sizeof(i));
- i.msg = "Battery low warning"; dictionary_set(known_journal_messages_ids, "e6f456bd92004d9580160b2207555186", &i, sizeof(i));
- i.msg = "Battery low poweroff"; dictionary_set(known_journal_messages_ids, "267437d33fdd41099ad76221cc24a335", &i, sizeof(i));
- i.msg = "Core mainloop failed"; dictionary_set(known_journal_messages_ids, "79e05b67bc4545d1922fe47107ee60c5", &i, sizeof(i));
- i.msg = "Core no xdgdir path"; dictionary_set(known_journal_messages_ids, "dbb136b10ef4457ba47a795d62f108c9", &i, sizeof(i));
- i.msg = "Core capability bounding user"; dictionary_set(known_journal_messages_ids, "ed158c2df8884fa584eead2d902c1032", &i, sizeof(i));
- i.msg = "Core capability bounding"; dictionary_set(known_journal_messages_ids, "42695b500df048298bee37159caa9f2e", &i, sizeof(i));
- i.msg = "Core disable privileges"; dictionary_set(known_journal_messages_ids, "bfc2430724ab44499735b4f94cca9295", &i, sizeof(i));
- i.msg = "Core start target failed"; dictionary_set(known_journal_messages_ids, "59288af523be43a28d494e41e26e4510", &i, sizeof(i));
- i.msg = "Core isolate target failed"; dictionary_set(known_journal_messages_ids, "689b4fcc97b4486ea5da92db69c9e314", &i, sizeof(i));
- i.msg = "Core fd set failed"; dictionary_set(known_journal_messages_ids, "5ed836f1766f4a8a9fc5da45aae23b29", &i, sizeof(i));
- i.msg = "Core pid1 environment"; dictionary_set(known_journal_messages_ids, "6a40fbfbd2ba4b8db02fb40c9cd090d7", &i, sizeof(i));
- i.msg = "Core manager allocate"; dictionary_set(known_journal_messages_ids, "0e54470984ac419689743d957a119e2e", &i, sizeof(i));
- i.msg = "Smack failed write"; dictionary_set(known_journal_messages_ids, "d67fa9f847aa4b048a2ae33535331adb", &i, sizeof(i));
- i.msg = "Shutdown error"; dictionary_set(known_journal_messages_ids, "af55a6f75b544431b72649f36ff6d62c", &i, sizeof(i));
- i.msg = "Valgrind helper fork"; dictionary_set(known_journal_messages_ids, "d18e0339efb24a068d9c1060221048c2", &i, sizeof(i));
- i.msg = "Unit starting"; dictionary_set(known_journal_messages_ids, "7d4958e842da4a758f6c1cdc7b36dcc5", &i, sizeof(i));
- i.msg = "Unit started"; dictionary_set(known_journal_messages_ids, "39f53479d3a045ac8e11786248231fbf", &i, sizeof(i));
- i.msg = "Unit failed"; dictionary_set(known_journal_messages_ids, "be02cf6855d2428ba40df7e9d022f03d", &i, sizeof(i));
- i.msg = "Unit stopping"; dictionary_set(known_journal_messages_ids, "de5b426a63be47a7b6ac3eaac82e2f6f", &i, sizeof(i));
- i.msg = "Unit stopped"; dictionary_set(known_journal_messages_ids, "9d1aaa27d60140bd96365438aad20286", &i, sizeof(i));
- i.msg = "Unit reloading"; dictionary_set(known_journal_messages_ids, "d34d037fff1847e6ae669a370e694725", &i, sizeof(i));
- i.msg = "Unit reloaded"; dictionary_set(known_journal_messages_ids, "7b05ebc668384222baa8881179cfda54", &i, sizeof(i));
- i.msg = "Unit restart scheduled"; dictionary_set(known_journal_messages_ids, "5eb03494b6584870a536b337290809b3", &i, sizeof(i));
- i.msg = "Unit resources"; dictionary_set(known_journal_messages_ids, "ae8f7b866b0347b9af31fe1c80b127c0", &i, sizeof(i));
- i.msg = "Unit success"; dictionary_set(known_journal_messages_ids, "7ad2d189f7e94e70a38c781354912448", &i, sizeof(i));
- i.msg = "Unit skipped"; dictionary_set(known_journal_messages_ids, "0e4284a0caca4bfc81c0bb6786972673", &i, sizeof(i));
- i.msg = "Unit failure result"; dictionary_set(known_journal_messages_ids, "d9b373ed55a64feb8242e02dbe79a49c", &i, sizeof(i));
- i.msg = "Spawn failed"; dictionary_set(known_journal_messages_ids, "641257651c1b4ec9a8624d7a40a9e1e7", &i, sizeof(i));
- i.msg = "Unit process exit"; dictionary_set(known_journal_messages_ids, "98e322203f7a4ed290d09fe03c09fe15", &i, sizeof(i));
- i.msg = "Forward syslog missed"; dictionary_set(known_journal_messages_ids, "0027229ca0644181a76c4e92458afa2e", &i, sizeof(i));
- i.msg = "Overmounting"; dictionary_set(known_journal_messages_ids, "1dee0369c7fc4736b7099b38ecb46ee7", &i, sizeof(i));
- i.msg = "Unit oomd kill"; dictionary_set(known_journal_messages_ids, "d989611b15e44c9dbf31e3c81256e4ed", &i, sizeof(i));
- i.msg = "Unit out of memory"; dictionary_set(known_journal_messages_ids, "fe6faa94e7774663a0da52717891d8ef", &i, sizeof(i));
- i.msg = "Lid opened"; dictionary_set(known_journal_messages_ids, "b72ea4a2881545a0b50e200e55b9b06f", &i, sizeof(i));
- i.msg = "Lid closed"; dictionary_set(known_journal_messages_ids, "b72ea4a2881545a0b50e200e55b9b070", &i, sizeof(i));
- i.msg = "System docked"; dictionary_set(known_journal_messages_ids, "f5f416b862074b28927a48c3ba7d51ff", &i, sizeof(i));
- i.msg = "System undocked"; dictionary_set(known_journal_messages_ids, "51e171bd585248568110144c517cca53", &i, sizeof(i));
- i.msg = "Power key"; dictionary_set(known_journal_messages_ids, "b72ea4a2881545a0b50e200e55b9b071", &i, sizeof(i));
- i.msg = "Power key long press"; dictionary_set(known_journal_messages_ids, "3e0117101eb243c1b9a50db3494ab10b", &i, sizeof(i));
- i.msg = "Reboot key"; dictionary_set(known_journal_messages_ids, "9fa9d2c012134ec385451ffe316f97d0", &i, sizeof(i));
- i.msg = "Reboot key long press"; dictionary_set(known_journal_messages_ids, "f1c59a58c9d943668965c337caec5975", &i, sizeof(i));
- i.msg = "Suspend key"; dictionary_set(known_journal_messages_ids, "b72ea4a2881545a0b50e200e55b9b072", &i, sizeof(i));
- i.msg = "Suspend key long press"; dictionary_set(known_journal_messages_ids, "bfdaf6d312ab4007bc1fe40a15df78e8", &i, sizeof(i));
- i.msg = "Hibernate key"; dictionary_set(known_journal_messages_ids, "b72ea4a2881545a0b50e200e55b9b073", &i, sizeof(i));
- i.msg = "Hibernate key long press"; dictionary_set(known_journal_messages_ids, "167836df6f7f428e98147227b2dc8945", &i, sizeof(i));
- i.msg = "Invalid configuration"; dictionary_set(known_journal_messages_ids, "c772d24e9a884cbeb9ea12625c306c01", &i, sizeof(i));
- i.msg = "Dnssec failure"; dictionary_set(known_journal_messages_ids, "1675d7f172174098b1108bf8c7dc8f5d", &i, sizeof(i));
- i.msg = "Dnssec trust anchor revoked"; dictionary_set(known_journal_messages_ids, "4d4408cfd0d144859184d1e65d7c8a65", &i, sizeof(i));
- i.msg = "Dnssec downgrade"; dictionary_set(known_journal_messages_ids, "36db2dfa5a9045e1bd4af5f93e1cf057", &i, sizeof(i));
- i.msg = "Unsafe user name"; dictionary_set(known_journal_messages_ids, "b61fdac612e94b9182285b998843061f", &i, sizeof(i));
- i.msg = "Mount point path not suitable"; dictionary_set(known_journal_messages_ids, "1b3bb94037f04bbf81028e135a12d293", &i, sizeof(i));
- i.msg = "Device path not suitable"; dictionary_set(known_journal_messages_ids, "010190138f494e29a0ef6669749531aa", &i, sizeof(i));
- i.msg = "Nobody user unsuitable"; dictionary_set(known_journal_messages_ids, "b480325f9c394a7b802c231e51a2752c", &i, sizeof(i));
- i.msg = "Systemd udev settle deprecated"; dictionary_set(known_journal_messages_ids, "1c0454c1bd2241e0ac6fefb4bc631433", &i, sizeof(i));
- i.msg = "Time sync"; dictionary_set(known_journal_messages_ids, "7c8a41f37b764941a0e1780b1be2f037", &i, sizeof(i));
- i.msg = "Time bump"; dictionary_set(known_journal_messages_ids, "7db73c8af0d94eeb822ae04323fe6ab6", &i, sizeof(i));
- i.msg = "Shutdown scheduled"; dictionary_set(known_journal_messages_ids, "9e7066279dc8403da79ce4b1a69064b2", &i, sizeof(i));
- i.msg = "Shutdown canceled"; dictionary_set(known_journal_messages_ids, "249f6fb9e6e2428c96f3f0875681ffa3", &i, sizeof(i));
- i.msg = "TPM pcr extend"; dictionary_set(known_journal_messages_ids, "3f7d5ef3e54f4302b4f0b143bb270cab", &i, sizeof(i));
- i.msg = "Memory trim"; dictionary_set(known_journal_messages_ids, "f9b0be465ad540d0850ad32172d57c21", &i, sizeof(i));
- i.msg = "Sysv generator deprecated"; dictionary_set(known_journal_messages_ids, "a8fa8dacdb1d443e9503b8be367a6adb", &i, sizeof(i));
-
- // gnome
- // https://gitlab.gnome.org/GNOME/gnome-session/-/blob/main/gnome-session/gsm-manager.c
- i.msg = "Gnome SM startup succeeded"; dictionary_set(known_journal_messages_ids, "0ce153587afa4095832d233c17a88001", &i, sizeof(i));
- i.msg = "Gnome SM unrecoverable failure"; dictionary_set(known_journal_messages_ids, "10dd2dc188b54a5e98970f56499d1f73", &i, sizeof(i));
-
- // gnome-shell
- // https://gitlab.gnome.org/GNOME/gnome-shell/-/blob/main/js/ui/main.js#L56
- i.msg = "Gnome shell started";dictionary_set(known_journal_messages_ids, "f3ea493c22934e26811cd62abe8e203a", &i, sizeof(i));
-
- // flathub
- // https://docs.flatpak.org/de/latest/flatpak-command-reference.html
- i.msg = "Flatpak cache"; dictionary_set(known_journal_messages_ids, "c7b39b1e006b464599465e105b361485", &i, sizeof(i));
-
- // ???
- i.msg = "Flathub pulls"; dictionary_set(known_journal_messages_ids, "75ba3deb0af041a9a46272ff85d9e73e", &i, sizeof(i));
- i.msg = "Flathub pull errors"; dictionary_set(known_journal_messages_ids, "f02bce89a54e4efab3a94a797d26204a", &i, sizeof(i));
-
- // ??
- i.msg = "Boltd starting"; dictionary_set(known_journal_messages_ids, "dd11929c788e48bdbb6276fb5f26b08a", &i, sizeof(i));
-
- // Netdata
- i.msg = "Netdata connection from child"; dictionary_set(known_journal_messages_ids, "ed4cdb8f1beb4ad3b57cb3cae2d162fa", &i, sizeof(i));
- i.msg = "Netdata connection to parent"; dictionary_set(known_journal_messages_ids, "6e2e3839067648968b646045dbf28d66", &i, sizeof(i));
- i.msg = "Netdata alert transition"; dictionary_set(known_journal_messages_ids, "9ce0cb58ab8b44df82c4bf1ad9ee22de", &i, sizeof(i));
- i.msg = "Netdata alert notification"; dictionary_set(known_journal_messages_ids, "6db0018e83e34320ae2a659d78019fb7", &i, sizeof(i));
-}
-
-void netdata_systemd_journal_transform_message_id(FACETS *facets __maybe_unused, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope __maybe_unused, void *data __maybe_unused) {
- const char *message_id = buffer_tostring(wb);
- struct message_id_info *i = dictionary_get(known_journal_messages_ids, message_id);
-
- if(!i)
- return;
-
- switch(scope) {
- default:
- case FACETS_TRANSFORM_DATA:
- case FACETS_TRANSFORM_VALUE:
- buffer_sprintf(wb, " (%s)", i->msg);
- break;
-
- case FACETS_TRANSFORM_FACET:
- case FACETS_TRANSFORM_FACET_SORT:
- case FACETS_TRANSFORM_HISTOGRAM:
- buffer_flush(wb);
- buffer_strcat(wb, i->msg);
- break;
- }
-}
-
-// ----------------------------------------------------------------------------
-
-static void netdata_systemd_journal_rich_message(FACETS *facets __maybe_unused, BUFFER *json_array, FACET_ROW_KEY_VALUE *rkv, FACET_ROW *row __maybe_unused, void *data __maybe_unused) {
- buffer_json_add_array_item_object(json_array);
- buffer_json_member_add_string(json_array, "value", buffer_tostring(rkv->wb));
- buffer_json_object_close(json_array);
-}
diff --git a/collectors/systemd-journal.plugin/systemd-main.c b/collectors/systemd-journal.plugin/systemd-main.c
deleted file mode 100644
index a3510b0ed..000000000
--- a/collectors/systemd-journal.plugin/systemd-main.c
+++ /dev/null
@@ -1,112 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-#include "systemd-internals.h"
-#include "libnetdata/required_dummies.h"
-
-#define SYSTEMD_JOURNAL_WORKER_THREADS 5
-
-netdata_mutex_t stdout_mutex = NETDATA_MUTEX_INITIALIZER;
-static bool plugin_should_exit = false;
-
-int main(int argc __maybe_unused, char **argv __maybe_unused) {
- clocks_init();
- netdata_thread_set_tag("SDMAIN");
- nd_log_initialize_for_external_plugins("systemd-journal.plugin");
-
- netdata_configured_host_prefix = getenv("NETDATA_HOST_PREFIX");
- if(verify_netdata_host_prefix(true) == -1) exit(1);
-
- // ------------------------------------------------------------------------
- // initialization
-
- netdata_systemd_journal_message_ids_init();
- journal_init_query_status();
- journal_init_files_and_directories();
-
- // ------------------------------------------------------------------------
- // debug
-
- if(argc == 2 && strcmp(argv[1], "debug") == 0) {
- journal_files_registry_update();
-
- bool cancelled = false;
- char buf[] = "systemd-journal after:-8640000 before:0 direction:backward last:200 data_only:false slice:true source:all";
- // char buf[] = "systemd-journal after:1695332964 before:1695937764 direction:backward last:100 slice:true source:all DHKucpqUoe1:PtVoyIuX.MU";
- // char buf[] = "systemd-journal after:1694511062 before:1694514662 anchor:1694514122024403";
- function_systemd_journal("123", buf, 600, &cancelled);
-// function_systemd_units("123", "systemd-units", 600, &cancelled);
- exit(1);
- }
-#ifdef ENABLE_SYSTEMD_DBUS
- if(argc == 2 && strcmp(argv[1], "debug-units") == 0) {
- bool cancelled = false;
- function_systemd_units("123", "systemd-units", 600, &cancelled);
- exit(1);
- }
-#endif
-
- // ------------------------------------------------------------------------
- // watcher thread
-
- netdata_thread_t watcher_thread;
- netdata_thread_create(&watcher_thread, "SDWATCH",
- NETDATA_THREAD_OPTION_DONT_LOG, journal_watcher_main, NULL);
-
- // ------------------------------------------------------------------------
- // the event loop for functions
-
- struct functions_evloop_globals *wg =
- functions_evloop_init(SYSTEMD_JOURNAL_WORKER_THREADS, "SDJ", &stdout_mutex, &plugin_should_exit);
-
- functions_evloop_add_function(wg, SYSTEMD_JOURNAL_FUNCTION_NAME, function_systemd_journal,
- SYSTEMD_JOURNAL_DEFAULT_TIMEOUT);
-
-#ifdef ENABLE_SYSTEMD_DBUS
- functions_evloop_add_function(wg, SYSTEMD_UNITS_FUNCTION_NAME, function_systemd_units,
- SYSTEMD_UNITS_DEFAULT_TIMEOUT);
-#endif
-
- // ------------------------------------------------------------------------
- // register functions to netdata
-
- netdata_mutex_lock(&stdout_mutex);
-
- fprintf(stdout, PLUGINSD_KEYWORD_FUNCTION " GLOBAL \"%s\" %d \"%s\"\n",
- SYSTEMD_JOURNAL_FUNCTION_NAME, SYSTEMD_JOURNAL_DEFAULT_TIMEOUT, SYSTEMD_JOURNAL_FUNCTION_DESCRIPTION);
-
-#ifdef ENABLE_SYSTEMD_DBUS
- fprintf(stdout, PLUGINSD_KEYWORD_FUNCTION " GLOBAL \"%s\" %d \"%s\"\n",
- SYSTEMD_UNITS_FUNCTION_NAME, SYSTEMD_UNITS_DEFAULT_TIMEOUT, SYSTEMD_UNITS_FUNCTION_DESCRIPTION);
-#endif
-
- fflush(stdout);
- netdata_mutex_unlock(&stdout_mutex);
-
- // ------------------------------------------------------------------------
-
- usec_t step_ut = 100 * USEC_PER_MS;
- usec_t send_newline_ut = 0;
- usec_t since_last_scan_ut = SYSTEMD_JOURNAL_ALL_FILES_SCAN_EVERY_USEC * 2; // something big to trigger scanning at start
- bool tty = isatty(fileno(stderr)) == 1;
-
- heartbeat_t hb;
- heartbeat_init(&hb);
- while(!plugin_should_exit) {
-
- if(since_last_scan_ut > SYSTEMD_JOURNAL_ALL_FILES_SCAN_EVERY_USEC) {
- journal_files_registry_update();
- since_last_scan_ut = 0;
- }
-
- usec_t dt_ut = heartbeat_next(&hb, step_ut);
- since_last_scan_ut += dt_ut;
- send_newline_ut += dt_ut;
-
- if(!tty && send_newline_ut > USEC_PER_SEC) {
- send_newline_and_flush();
- send_newline_ut = 0;
- }
- }
-
- exit(0);
-}
diff --git a/collectors/tc.plugin/Makefile.am b/collectors/tc.plugin/Makefile.am
deleted file mode 100644
index 9bc6cf2de..000000000
--- a/collectors/tc.plugin/Makefile.am
+++ /dev/null
@@ -1,20 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-CLEANFILES = \
- tc-qos-helper.sh \
- $(NULL)
-
-include $(top_srcdir)/build/subst.inc
-SUFFIXES = .in
-
-dist_plugins_SCRIPTS = \
- tc-qos-helper.sh \
- $(NULL)
-
-dist_noinst_DATA = \
- tc-qos-helper.sh.in \
- README.md \
- $(NULL)
diff --git a/collectors/timex.plugin/Makefile.am b/collectors/timex.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/timex.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/timex.plugin/metadata.yaml b/collectors/timex.plugin/metadata.yaml
deleted file mode 100644
index 2b43d8a24..000000000
--- a/collectors/timex.plugin/metadata.yaml
+++ /dev/null
@@ -1,112 +0,0 @@
-plugin_name: timex.plugin
-modules:
- - meta:
- plugin_name: timex.plugin
- module_name: timex.plugin
- monitored_instance:
- name: Timex
- link: ""
- categories:
- - data-collection.system-clock-and-ntp
- icon_filename: "syslog.png"
- related_resources:
- integrations:
- list: []
- info_provided_to_referring_integrations:
- description: ""
- keywords: []
- most_popular: false
- overview:
- data_collection:
- metrics_description: "Examine Timex metrics to gain insights into system clock operations. Study time sync status, clock drift, and adjustments to ensure accurate system timekeeping."
- method_description: "It uses system call adjtimex on Linux and ntp_adjtime on FreeBSD or Mac to monitor the system kernel clock synchronization state."
- supported_platforms:
- include: []
- exclude: []
- multi_instance: true
- additional_permissions:
- description: ""
- default_behavior:
- auto_detection:
- description: ""
- limits:
- description: ""
- performance_impact:
- description: ""
- setup:
- prerequisites:
- list: []
- configuration:
- file:
- name: "netdata.conf"
- section_name: "[plugin:timex]"
- description: "The netdata main configuration file."
- options:
- description: "At least one option ('clock synchronization state', 'time offset') needs to be enabled for this collector to run."
- folding:
- title: "Config options"
- enabled: true
- list:
- - name: update every
- description: Data collection frequency.
- default_value: 1
- required: false
- - name: clock synchronization state
- description: Make chart showing system clock synchronization state.
- default_value: yes
- required: true
- - name: time offset
- description: Make chart showing computed time offset between local system and reference clock
- default_value: yes
- required: true
- examples:
- folding:
- enabled: true
- title: "Config"
- list:
- - name: Basic
- description: A basic configuration example.
- config: |
- [plugin:timex]
- update every = 1
- clock synchronization state = yes
- time offset = yes
- troubleshooting:
- problems:
- list: []
- alerts:
- - name: system_clock_sync_state
- link: https://github.com/netdata/netdata/blob/master/health/health.d/timex.conf
- metric: system.clock_sync_state
- info: when set to 0, the system kernel believes the system clock is not properly synchronized to a reliable server
- os: "linux"
- metrics:
- folding:
- title: Metrics
- enabled: false
- description: ""
- availability: []
- scopes:
- - name: global
- description: "These metrics refer to the entire monitored application."
- labels: []
- metrics:
- - name: system.clock_sync_state
- description: System Clock Synchronization State
- unit: "state"
- chart_type: line
- dimensions:
- - name: state
- - name: system.clock_status
- description: System Clock Status
- unit: "status"
- chart_type: line
- dimensions:
- - name: unsync
- - name: clockerr
- - name: system.clock_sync_offset
- description: Computed Time Offset Between Local System and Reference Clock
- unit: "milliseconds"
- chart_type: line
- dimensions:
- - name: offset
diff --git a/collectors/xenstat.plugin/Makefile.am b/collectors/xenstat.plugin/Makefile.am
deleted file mode 100644
index 161784b8f..000000000
--- a/collectors/xenstat.plugin/Makefile.am
+++ /dev/null
@@ -1,8 +0,0 @@
-# SPDX-License-Identifier: GPL-3.0-or-later
-
-AUTOMAKE_OPTIONS = subdir-objects
-MAINTAINERCLEANFILES = $(srcdir)/Makefile.in
-
-dist_noinst_DATA = \
- README.md \
- $(NULL)
diff --git a/collectors/systemd-journal.plugin/active_journal_centralization_guide_no_encryption.md b/docs/observability-centralization-points/logs-centralization-points-with-systemd-journald/active-journal-source-without-encryption.md
index cbed1e81e..cbed1e81e 100644
--- a/collectors/systemd-journal.plugin/active_journal_centralization_guide_no_encryption.md
+++ b/docs/observability-centralization-points/logs-centralization-points-with-systemd-journald/active-journal-source-without-encryption.md
diff --git a/collectors/systemd-journal.plugin/passive_journal_centralization_guide_no_encryption.md b/docs/observability-centralization-points/logs-centralization-points-with-systemd-journald/passive-journal-centralization-without-encryption.md
index b70c22033..b70c22033 100644
--- a/collectors/systemd-journal.plugin/passive_journal_centralization_guide_no_encryption.md
+++ b/docs/observability-centralization-points/logs-centralization-points-with-systemd-journald/passive-journal-centralization-without-encryption.md
diff --git a/collectors/all.h b/src/collectors/all.h
index 38241dfa9..24236e9bf 100644
--- a/collectors/all.h
+++ b/src/collectors/all.h
@@ -59,16 +59,14 @@
// CPU per core
#define NETDATA_CHART_PRIO_CPU_PER_CORE 1000 // +1 per core
+#define NETDATA_CHART_PRIO_CPUFREQ_SCALING_CUR_FREQ 1002
+#define NETDATA_CHART_PRIO_CORE_THROTTLING 1003
+#define NETDATA_CHART_PRIO_PACKAGE_THROTTLING 1004
+#define NETDATA_CHART_PRIO_CPUIDLE 1005
+#define NETDATA_CHART_PRIO_INTERRUPTS_PER_CORE 1006 // +1 per core
+#define NETDATA_CHART_PRIO_POWERCAP 1007 // Linux powercap
#define NETDATA_CHART_PRIO_CPU_TEMPERATURE 1050 // freebsd only
-#define NETDATA_CHART_PRIO_CPUFREQ_SCALING_CUR_FREQ 5003 // freebsd only
-#define NETDATA_CHART_PRIO_CPUIDLE 6000
-#define NETDATA_CHART_PRIO_CORE_THROTTLING 5001
-#define NETDATA_CHART_PRIO_PACKAGE_THROTTLING 5002
-
-// Interrupts per core
-
-#define NETDATA_CHART_PRIO_INTERRUPTS_PER_CORE 1100 // +1 per core
// Memory Section - 1xxx
@@ -384,10 +382,6 @@
#define NETDATA_CHART_PRIO_POWER_SUPPLY_ENERGY 9502
#define NETDATA_CHART_PRIO_POWER_SUPPLY_VOLTAGE 9503
-// Linux powercap
-
-#define NETDATA_CHART_PRIO_POWERCAP 9600
-
// Wireless
#define NETDATA_CHART_PRIO_WIRELESS_IFACE 7110
diff --git a/collectors/apps.plugin/apps_groups.conf b/src/collectors/apps.plugin/apps_groups.conf
index 195536a0a..b6d11bbf8 100644
--- a/collectors/apps.plugin/apps_groups.conf
+++ b/src/collectors/apps.plugin/apps_groups.conf
@@ -84,6 +84,7 @@ perf.plugin: perf.plugin
charts.d.plugin: *charts.d.plugin*
python.d.plugin: *python.d.plugin*
systemd-journal.plugin:*systemd-journal.plugin*
+network-viewer.plugin:*network-viewer.plugin*
tc-qos-helper: *tc-qos-helper.sh*
fping: fping
ioping: ioping
diff --git a/collectors/apps.plugin/integrations/applications.md b/src/collectors/apps.plugin/integrations/applications.md
index e5219fcc2..ed9a8a5df 100644
--- a/collectors/apps.plugin/integrations/applications.md
+++ b/src/collectors/apps.plugin/integrations/applications.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/apps.plugin/integrations/applications.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/apps.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/apps.plugin/integrations/applications.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/apps.plugin/metadata.yaml"
sidebar_label: "Applications"
learn_status: "Published"
-learn_rel_path: "Data Collection/Processes and System Services"
+learn_rel_path: "Collecting Metrics/Processes and System Services"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/apps.plugin/integrations/user_groups.md b/src/collectors/apps.plugin/integrations/user_groups.md
index 4ccbfc95f..5a3939f1e 100644
--- a/collectors/apps.plugin/integrations/user_groups.md
+++ b/src/collectors/apps.plugin/integrations/user_groups.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/apps.plugin/integrations/user_groups.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/apps.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/apps.plugin/integrations/user_groups.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/apps.plugin/metadata.yaml"
sidebar_label: "User Groups"
learn_status: "Published"
-learn_rel_path: "Data Collection/Processes and System Services"
+learn_rel_path: "Collecting Metrics/Processes and System Services"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/apps.plugin/integrations/users.md b/src/collectors/apps.plugin/integrations/users.md
index c151fd8a2..83423ab73 100644
--- a/collectors/apps.plugin/integrations/users.md
+++ b/src/collectors/apps.plugin/integrations/users.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/apps.plugin/integrations/users.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/apps.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/apps.plugin/integrations/users.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/apps.plugin/metadata.yaml"
sidebar_label: "Users"
learn_status: "Published"
-learn_rel_path: "Data Collection/Processes and System Services"
+learn_rel_path: "Collecting Metrics/Processes and System Services"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/apps.plugin/metadata.yaml b/src/collectors/apps.plugin/metadata.yaml
index f5f22853a..f5f22853a 100644
--- a/collectors/apps.plugin/metadata.yaml
+++ b/src/collectors/apps.plugin/metadata.yaml
diff --git a/collectors/cgroups.plugin/cgroup-charts.c b/src/collectors/cgroups.plugin/cgroup-charts.c
index a89e8ac45..1a337cd99 100644
--- a/collectors/cgroups.plugin/cgroup-charts.c
+++ b/src/collectors/cgroups.plugin/cgroup-charts.c
@@ -96,7 +96,7 @@ void update_cpu_utilization_limit_chart(struct cgroup *cg, NETDATA_DOUBLE cpu_li
cg->prev_cpu_usage = cpu_usage;
- rrdsetvar_custom_chart_variable_set(cg->st_cpu, cg->chart_var_cpu_limit, cpu_limit);
+ rrdvar_chart_variable_set(cg->st_cpu, cg->chart_var_cpu_limit, cpu_limit);
rrdset_done(chart);
}
diff --git a/collectors/cgroups.plugin/cgroup-discovery.c b/src/collectors/cgroups.plugin/cgroup-discovery.c
index ede35ed8a..e5d029cfb 100644
--- a/collectors/cgroups.plugin/cgroup-discovery.c
+++ b/src/collectors/cgroups.plugin/cgroup-discovery.c
@@ -25,6 +25,10 @@ char cgroup_chart_id_prefix[] = "cgroup_";
char services_chart_id_prefix[] = "systemd_";
char *cgroups_rename_script = NULL;
+// Shared memory with information from detected cgroups
+netdata_ebpf_cgroup_shm_t shm_cgroup_ebpf = {NULL, NULL};
+int shm_fd_cgroup_ebpf = -1;
+sem_t *shm_mutex_cgroup_ebpf = SEM_FAILED;
// ----------------------------------------------------------------------------
@@ -42,7 +46,7 @@ static inline void cgroup_free_network_interfaces(struct cgroup *cg) {
cg->interfaces = i->next;
// delete the registration of proc_net_dev rename
- netdev_rename_device_del(i->host_device);
+ cgroup_rename_task_device_del(i->host_device);
freez((void *)i->host_device);
freez((void *)i->container_device);
@@ -1027,6 +1031,82 @@ static int discovery_is_cgroup_duplicate(struct cgroup *cg) {
}
// ----------------------------------------------------------------------------
+// ebpf shared memory
+
+static void netdata_cgroup_ebpf_set_values(size_t length)
+{
+ sem_wait(shm_mutex_cgroup_ebpf);
+
+ shm_cgroup_ebpf.header->cgroup_max = cgroup_root_max;
+ shm_cgroup_ebpf.header->systemd_enabled = cgroup_enable_systemd_services |
+ cgroup_enable_systemd_services_detailed_memory |
+ cgroup_used_memory;
+ shm_cgroup_ebpf.header->body_length = length;
+
+ sem_post(shm_mutex_cgroup_ebpf);
+}
+
+static void netdata_cgroup_ebpf_initialize_shm()
+{
+ shm_fd_cgroup_ebpf = shm_open(NETDATA_SHARED_MEMORY_EBPF_CGROUP_NAME, O_CREAT | O_RDWR, 0660);
+ if (shm_fd_cgroup_ebpf < 0) {
+ collector_error("Cannot initialize shared memory used by cgroup and eBPF, integration won't happen.");
+ return;
+ }
+
+ size_t length = sizeof(netdata_ebpf_cgroup_shm_header_t) + cgroup_root_max * sizeof(netdata_ebpf_cgroup_shm_body_t);
+ if (ftruncate(shm_fd_cgroup_ebpf, length)) {
+ collector_error("Cannot set size for shared memory.");
+ goto end_init_shm;
+ }
+
+ shm_cgroup_ebpf.header = (netdata_ebpf_cgroup_shm_header_t *) mmap(NULL, length,
+ PROT_READ | PROT_WRITE, MAP_SHARED,
+ shm_fd_cgroup_ebpf, 0);
+
+ if (unlikely(MAP_FAILED == shm_cgroup_ebpf.header)) {
+ shm_cgroup_ebpf.header = NULL;
+ collector_error("Cannot map shared memory used between cgroup and eBPF, integration won't happen");
+ goto end_init_shm;
+ }
+ shm_cgroup_ebpf.body = (netdata_ebpf_cgroup_shm_body_t *) ((char *)shm_cgroup_ebpf.header +
+ sizeof(netdata_ebpf_cgroup_shm_header_t));
+
+ shm_mutex_cgroup_ebpf = sem_open(NETDATA_NAMED_SEMAPHORE_EBPF_CGROUP_NAME, O_CREAT,
+ S_IRUSR | S_IWUSR | S_IRGRP | S_IWGRP | S_IROTH | S_IWOTH, 1);
+
+ if (shm_mutex_cgroup_ebpf != SEM_FAILED) {
+ netdata_cgroup_ebpf_set_values(length);
+ return;
+ }
+
+ collector_error("Cannot create semaphore, integration between eBPF and cgroup won't happen");
+ munmap(shm_cgroup_ebpf.header, length);
+ shm_cgroup_ebpf.header = NULL;
+
+ end_init_shm:
+ close(shm_fd_cgroup_ebpf);
+ shm_fd_cgroup_ebpf = -1;
+ shm_unlink(NETDATA_SHARED_MEMORY_EBPF_CGROUP_NAME);
+}
+
+static void cgroup_cleanup_ebpf_integration()
+{
+ if (shm_mutex_cgroup_ebpf != SEM_FAILED) {
+ sem_close(shm_mutex_cgroup_ebpf);
+ }
+
+ if (shm_cgroup_ebpf.header) {
+ shm_cgroup_ebpf.header->cgroup_root_count = 0;
+ munmap(shm_cgroup_ebpf.header, shm_cgroup_ebpf.header->body_length);
+ }
+
+ if (shm_fd_cgroup_ebpf > 0) {
+ close(shm_fd_cgroup_ebpf);
+ }
+}
+
+// ----------------------------------------------------------------------------
// cgroup network interfaces
#define CGROUP_NETWORK_INTERFACE_MAX_LINE 2048
@@ -1084,8 +1164,13 @@ static inline void read_cgroup_network_interfaces(struct cgroup *cg) {
collector_info("CGROUP: cgroup '%s' has network interface '%s' as '%s'", cg->id, i->host_device, i->container_device);
// register a device rename to proc_net_dev.c
- netdev_rename_device_add(i->host_device, i->container_device, cg->chart_id, cg->chart_labels,
- k8s_is_kubepod(cg) ? "k8s." : "", cgroup_netdev_get(cg));
+ cgroup_rename_task_add(
+ i->host_device,
+ i->container_device,
+ cg->chart_id,
+ cg->chart_labels,
+ k8s_is_kubepod(cg) ? "k8s." : "",
+ cgroup_netdev_get(cg));
}
}
@@ -1226,6 +1311,8 @@ void cgroup_discovery_worker(void *ptr)
service_register(SERVICE_THREAD_TYPE_LIBUV, NULL, NULL, NULL, false);
+ netdata_cgroup_ebpf_initialize_shm();
+
while (service_running(SERVICE_COLLECTORS)) {
worker_is_idle();
@@ -1239,6 +1326,7 @@ void cgroup_discovery_worker(void *ptr)
discovery_find_all_cgroups();
}
collector_info("discovery thread stopped");
+ cgroup_cleanup_ebpf_integration();
worker_unregister();
service_exits();
__atomic_store_n(&discovery_thread.exited,1,__ATOMIC_RELAXED);
diff --git a/collectors/cgroups.plugin/cgroup-internals.h b/src/collectors/cgroups.plugin/cgroup-internals.h
index a69802240..1f5be7707 100644
--- a/collectors/cgroups.plugin/cgroup-internals.h
+++ b/src/collectors/cgroups.plugin/cgroup-internals.h
@@ -287,16 +287,16 @@ struct cgroup {
char *filename_cpu_cfs_quota;
unsigned long long cpu_cfs_quota;
- const RRDSETVAR_ACQUIRED *chart_var_cpu_limit;
+ const RRDVAR_ACQUIRED *chart_var_cpu_limit;
NETDATA_DOUBLE prev_cpu_usage;
char *filename_memory_limit;
unsigned long long memory_limit;
- const RRDSETVAR_ACQUIRED *chart_var_memory_limit;
+ const RRDVAR_ACQUIRED *chart_var_memory_limit;
char *filename_memoryswap_limit;
unsigned long long memoryswap_limit;
- const RRDSETVAR_ACQUIRED *chart_var_memoryswap_limit;
+ const RRDVAR_ACQUIRED *chart_var_memoryswap_limit;
const DICTIONARY_ITEM *cgroup_netdev_link;
@@ -386,10 +386,6 @@ extern uint32_t throttled_time_hash;
extern uint32_t throttled_usec_hash;
extern struct cgroup *cgroup_root;
-extern netdata_ebpf_cgroup_shm_t shm_cgroup_ebpf;
-extern int shm_fd_cgroup_ebpf;
-extern sem_t *shm_mutex_cgroup_ebpf;
-
enum cgroups_type { CGROUPS_AUTODETECT_FAIL, CGROUPS_V1, CGROUPS_V2 };
enum cgroups_systemd_setting {
@@ -452,15 +448,10 @@ static inline char *cgroup_chart_type(char *buffer, struct cgroup *cg) {
}
#define RRDFUNCTIONS_CGTOP_HELP "View running containers"
+#define RRDFUNCTIONS_SYSTEMD_SERVICES_HELP "View systemd services"
-int cgroup_function_cgroup_top(BUFFER *wb, int timeout, const char *function, void *collector_data,
- rrd_function_result_callback_t result_cb, void *result_cb_data,
- rrd_function_is_cancelled_cb_t is_cancelled_cb, void *is_cancelled_cb_data,
- rrd_function_register_canceller_cb_t register_canceller_cb, void *register_canceller_cb_data);
-int cgroup_function_systemd_top(BUFFER *wb, int timeout, const char *function, void *collector_data,
- rrd_function_result_callback_t result_cb, void *result_cb_data,
- rrd_function_is_cancelled_cb_t is_cancelled_cb, void *is_cancelled_cb_data,
- rrd_function_register_canceller_cb_t register_canceller_cb, void *register_canceller_cb_data);
+int cgroup_function_cgroup_top(BUFFER *wb, const char *function);
+int cgroup_function_systemd_top(BUFFER *wb, const char *function);
void cgroup_netdev_link_init(void);
const DICTIONARY_ITEM *cgroup_netdev_get(struct cgroup *cg);
diff --git a/collectors/cgroups.plugin/cgroup-name.sh.in b/src/collectors/cgroups.plugin/cgroup-name.sh.in
index 0f8b63256..0f8b63256 100755
--- a/collectors/cgroups.plugin/cgroup-name.sh.in
+++ b/src/collectors/cgroups.plugin/cgroup-name.sh.in
diff --git a/collectors/cgroups.plugin/cgroup-network-helper.sh.in b/src/collectors/cgroups.plugin/cgroup-network-helper.sh.in
index da9b9162a..da9b9162a 100755
--- a/collectors/cgroups.plugin/cgroup-network-helper.sh.in
+++ b/src/collectors/cgroups.plugin/cgroup-network-helper.sh.in
diff --git a/collectors/cgroups.plugin/cgroup-network.c b/src/collectors/cgroups.plugin/cgroup-network.c
index 508ea07c6..085a6aa6f 100644
--- a/collectors/cgroups.plugin/cgroup-network.c
+++ b/src/collectors/cgroups.plugin/cgroup-network.c
@@ -7,7 +7,6 @@
#ifndef _GNU_SOURCE
#define _GNU_SOURCE /* See feature_test_macros(7) */
#endif
-#include <sched.h>
#endif
char env_netdata_host_prefix[FILENAME_MAX + 50] = "";
@@ -183,7 +182,7 @@ int proc_pid_fd(const char *prefix, const char *ns, pid_t pid) {
char filename[FILENAME_MAX + 1];
snprintfz(filename, FILENAME_MAX, "%s/proc/%d/%s", prefix, (int)pid, ns);
- int fd = open(filename, O_RDONLY);
+ int fd = open(filename, O_RDONLY | O_CLOEXEC);
if(fd == -1)
collector_error("Cannot open proc_pid_fd() file '%s'", filename);
diff --git a/collectors/cgroups.plugin/cgroup-top.c b/src/collectors/cgroups.plugin/cgroup-top.c
index 8d44d3b56..aa413dad1 100644
--- a/collectors/cgroups.plugin/cgroup-top.c
+++ b/src/collectors/cgroups.plugin/cgroup-top.c
@@ -98,13 +98,7 @@ void cgroup_netdev_get_bandwidth(struct cgroup *cg, NETDATA_DOUBLE *received, NE
*sent = t->sent[slot];
}
-int cgroup_function_cgroup_top(BUFFER *wb, int timeout __maybe_unused, const char *function __maybe_unused,
- void *collector_data __maybe_unused,
- rrd_function_result_callback_t result_cb, void *result_cb_data,
- rrd_function_is_cancelled_cb_t is_cancelled_cb, void *is_cancelled_cb_data,
- rrd_function_register_canceller_cb_t register_canceller_cb __maybe_unused,
- void *register_canceller_cb_data __maybe_unused) {
-
+int cgroup_function_cgroup_top(BUFFER *wb, const char *function __maybe_unused) {
buffer_flush(wb);
wb->content_type = CT_APPLICATION_JSON;
buffer_json_initialize(wb, "\"", "\"", 0, true, BUFFER_JSON_OPTIONS_DEFAULT);
@@ -113,11 +107,13 @@ int cgroup_function_cgroup_top(BUFFER *wb, int timeout __maybe_unused, const cha
buffer_json_member_add_uint64(wb, "status", HTTP_RESP_OK);
buffer_json_member_add_string(wb, "type", "table");
buffer_json_member_add_time_t(wb, "update_every", 1);
+ buffer_json_member_add_boolean(wb, "has_history", false);
buffer_json_member_add_string(wb, "help", RRDFUNCTIONS_CGTOP_HELP);
buffer_json_member_add_array(wb, "data");
double max_pids = 0.0;
double max_cpu = 0.0;
+ double max_cpu_throttled = 0.0;
double max_ram = 0.0;
double max_disk_io_read = 0.0;
double max_disk_io_written = 0.0;
@@ -149,6 +145,9 @@ int cgroup_function_cgroup_top(BUFFER *wb, int timeout __maybe_unused, const cha
max_cpu = MAX(max_cpu, cpu);
}
+ double cpu_throttled = (double)cg->cpuacct_cpu_throttling.nr_throttled_perc;
+ max_cpu_throttled = MAX(max_cpu_throttled, cpu_throttled);
+
double ram = rrddim_get_last_stored_value(cg->st_mem_rd_ram, &max_ram, 1.0);
rd = cg->st_throttle_io_rd_read ? cg->st_throttle_io_rd_read : cg->st_io_rd_read;
@@ -167,6 +166,7 @@ int cgroup_function_cgroup_top(BUFFER *wb, int timeout __maybe_unused, const cha
buffer_json_add_array_item_double(wb, pids_current);
buffer_json_add_array_item_double(wb, cpu);
+ buffer_json_add_array_item_double(wb, cpu_throttled);
buffer_json_add_array_item_double(wb, ram);
buffer_json_add_array_item_double(wb, disk_io_read);
buffer_json_add_array_item_double(wb, disk_io_written);
@@ -215,6 +215,13 @@ int cgroup_function_cgroup_top(BUFFER *wb, int timeout __maybe_unused, const cha
RRDF_FIELD_OPTS_VISIBLE,
NULL);
+ buffer_rrdf_table_add_field(wb, field_id++, "CPU Throttling", "CPU Throttled Runnable Periods",
+ RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER,
+ 0, "%", max_cpu_throttled, RRDF_FIELD_SORT_DESCENDING, NULL,
+ RRDF_FIELD_SUMMARY_SUM, RRDF_FIELD_FILTER_NONE,
+ is_inside_k8s ? RRDF_FIELD_OPTS_VISIBLE : RRDF_FIELD_OPTS_NONE,
+ NULL);
+
// RAM
buffer_rrdf_table_add_field(wb, field_id++, "RAM", "RAM Usage",
RRDF_FIELD_TYPE_BAR_WITH_INTEGER, RRDF_FIELD_VISUAL_BAR, RRDF_FIELD_TRANSFORM_NUMBER,
@@ -331,25 +338,10 @@ int cgroup_function_cgroup_top(BUFFER *wb, int timeout __maybe_unused, const cha
buffer_json_member_add_time_t(wb, "expires", now_realtime_sec() + 1);
buffer_json_finalize(wb);
- int response = HTTP_RESP_OK;
- if(is_cancelled_cb && is_cancelled_cb(is_cancelled_cb_data)) {
- buffer_flush(wb);
- response = HTTP_RESP_CLIENT_CLOSED_REQUEST;
- }
-
- if(result_cb)
- result_cb(wb, response, result_cb_data);
-
- return response;
+ return HTTP_RESP_OK;
}
-int cgroup_function_systemd_top(BUFFER *wb, int timeout __maybe_unused, const char *function __maybe_unused,
- void *collector_data __maybe_unused,
- rrd_function_result_callback_t result_cb, void *result_cb_data,
- rrd_function_is_cancelled_cb_t is_cancelled_cb, void *is_cancelled_cb_data,
- rrd_function_register_canceller_cb_t register_canceller_cb __maybe_unused,
- void *register_canceller_cb_data __maybe_unused) {
-
+int cgroup_function_systemd_top(BUFFER *wb, const char *function __maybe_unused) {
buffer_flush(wb);
wb->content_type = CT_APPLICATION_JSON;
buffer_json_initialize(wb, "\"", "\"", 0, true, BUFFER_JSON_OPTIONS_DEFAULT);
@@ -358,6 +350,7 @@ int cgroup_function_systemd_top(BUFFER *wb, int timeout __maybe_unused, const ch
buffer_json_member_add_uint64(wb, "status", HTTP_RESP_OK);
buffer_json_member_add_string(wb, "type", "table");
buffer_json_member_add_time_t(wb, "update_every", 1);
+ buffer_json_member_add_boolean(wb, "has_history", false);
buffer_json_member_add_string(wb, "help", RRDFUNCTIONS_CGTOP_HELP);
buffer_json_member_add_array(wb, "data");
@@ -507,14 +500,5 @@ int cgroup_function_systemd_top(BUFFER *wb, int timeout __maybe_unused, const ch
buffer_json_member_add_time_t(wb, "expires", now_realtime_sec() + 1);
buffer_json_finalize(wb);
- int response = HTTP_RESP_OK;
- if(is_cancelled_cb && is_cancelled_cb(is_cancelled_cb_data)) {
- buffer_flush(wb);
- response = HTTP_RESP_CLIENT_CLOSED_REQUEST;
- }
-
- if(result_cb)
- result_cb(wb, response, result_cb_data);
-
- return response;
+ return HTTP_RESP_OK;
}
diff --git a/collectors/cgroups.plugin/integrations/containers.md b/src/collectors/cgroups.plugin/integrations/containers.md
index 6273d1e91..e769fc5f9 100644
--- a/collectors/cgroups.plugin/integrations/containers.md
+++ b/src/collectors/cgroups.plugin/integrations/containers.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/integrations/containers.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/integrations/containers.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/metadata.yaml"
sidebar_label: "Containers"
learn_status: "Published"
-learn_rel_path: "Data Collection/Containers and VMs"
+learn_rel_path: "Collecting Metrics/Containers and VMs"
most_popular: True
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -140,10 +140,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
-| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
-| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
-| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
+| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
+| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
+| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
+| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
## Setup
diff --git a/collectors/cgroups.plugin/integrations/kubernetes_containers.md b/src/collectors/cgroups.plugin/integrations/kubernetes_containers.md
index 9be32a12a..c71b5736c 100644
--- a/collectors/cgroups.plugin/integrations/kubernetes_containers.md
+++ b/src/collectors/cgroups.plugin/integrations/kubernetes_containers.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/integrations/kubernetes_containers.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/integrations/kubernetes_containers.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/metadata.yaml"
sidebar_label: "Kubernetes Containers"
learn_status: "Published"
-learn_rel_path: "Data Collection/Kubernetes"
+learn_rel_path: "Collecting Metrics/Kubernetes"
most_popular: True
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -154,10 +154,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ k8s_cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | k8s.cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
-| [ k8s_cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | k8s.cgroup.mem_usage | cgroup memory utilization |
-| [ k8s_cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | k8s.cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
-| [ k8s_cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | k8s.cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
+| [ k8s_cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | k8s.cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
+| [ k8s_cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | k8s.cgroup.mem_usage | cgroup memory utilization |
+| [ k8s_cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | k8s.cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
+| [ k8s_cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | k8s.cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
## Setup
diff --git a/collectors/cgroups.plugin/integrations/libvirt_containers.md b/src/collectors/cgroups.plugin/integrations/libvirt_containers.md
index fed454698..409b43609 100644
--- a/collectors/cgroups.plugin/integrations/libvirt_containers.md
+++ b/src/collectors/cgroups.plugin/integrations/libvirt_containers.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/integrations/libvirt_containers.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/integrations/libvirt_containers.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/metadata.yaml"
sidebar_label: "Libvirt Containers"
learn_status: "Published"
-learn_rel_path: "Data Collection/Containers and VMs"
+learn_rel_path: "Collecting Metrics/Containers and VMs"
most_popular: True
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -140,10 +140,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
-| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
-| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
-| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
+| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
+| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
+| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
+| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
## Setup
diff --git a/collectors/cgroups.plugin/integrations/lxc_containers.md b/src/collectors/cgroups.plugin/integrations/lxc_containers.md
index 3f05ffd5f..14897e468 100644
--- a/collectors/cgroups.plugin/integrations/lxc_containers.md
+++ b/src/collectors/cgroups.plugin/integrations/lxc_containers.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/integrations/lxc_containers.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/integrations/lxc_containers.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/metadata.yaml"
sidebar_label: "LXC Containers"
learn_status: "Published"
-learn_rel_path: "Data Collection/Containers and VMs"
+learn_rel_path: "Collecting Metrics/Containers and VMs"
most_popular: True
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -140,10 +140,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
-| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
-| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
-| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
+| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
+| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
+| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
+| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
## Setup
diff --git a/collectors/cgroups.plugin/integrations/ovirt_containers.md b/src/collectors/cgroups.plugin/integrations/ovirt_containers.md
index 5771aeea1..49f8c8091 100644
--- a/collectors/cgroups.plugin/integrations/ovirt_containers.md
+++ b/src/collectors/cgroups.plugin/integrations/ovirt_containers.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/integrations/ovirt_containers.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/integrations/ovirt_containers.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/metadata.yaml"
sidebar_label: "oVirt Containers"
learn_status: "Published"
-learn_rel_path: "Data Collection/Containers and VMs"
+learn_rel_path: "Collecting Metrics/Containers and VMs"
most_popular: True
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -140,10 +140,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
-| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
-| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
-| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
+| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
+| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
+| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
+| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
## Setup
diff --git a/collectors/cgroups.plugin/integrations/proxmox_containers.md b/src/collectors/cgroups.plugin/integrations/proxmox_containers.md
index 1804a40ca..fa2177b46 100644
--- a/collectors/cgroups.plugin/integrations/proxmox_containers.md
+++ b/src/collectors/cgroups.plugin/integrations/proxmox_containers.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/integrations/proxmox_containers.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/integrations/proxmox_containers.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/metadata.yaml"
sidebar_label: "Proxmox Containers"
learn_status: "Published"
-learn_rel_path: "Data Collection/Containers and VMs"
+learn_rel_path: "Collecting Metrics/Containers and VMs"
most_popular: True
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -140,10 +140,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
-| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
-| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
-| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
+| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
+| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
+| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
+| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
## Setup
diff --git a/collectors/cgroups.plugin/integrations/systemd_services.md b/src/collectors/cgroups.plugin/integrations/systemd_services.md
index 0ce906366..e3a80d549 100644
--- a/collectors/cgroups.plugin/integrations/systemd_services.md
+++ b/src/collectors/cgroups.plugin/integrations/systemd_services.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/integrations/systemd_services.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/integrations/systemd_services.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/metadata.yaml"
sidebar_label: "Systemd Services"
learn_status: "Published"
-learn_rel_path: "Data Collection/Systemd"
+learn_rel_path: "Collecting Metrics/Systemd"
most_popular: True
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/cgroups.plugin/integrations/virtual_machines.md b/src/collectors/cgroups.plugin/integrations/virtual_machines.md
index 6a64923c4..b81c82fb9 100644
--- a/collectors/cgroups.plugin/integrations/virtual_machines.md
+++ b/src/collectors/cgroups.plugin/integrations/virtual_machines.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/integrations/virtual_machines.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/cgroups.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/integrations/virtual_machines.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/cgroups.plugin/metadata.yaml"
sidebar_label: "Virtual Machines"
learn_status: "Published"
-learn_rel_path: "Data Collection/Containers and VMs"
+learn_rel_path: "Collecting Metrics/Containers and VMs"
most_popular: True
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -140,10 +140,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
-| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
-| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
-| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
+| [ cgroup_10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.cpu_limit | average cgroup CPU utilization over the last 10 minutes |
+| [ cgroup_ram_in_use ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.mem_usage | cgroup memory utilization |
+| [ cgroup_1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | average number of packets received by the network interface ${label:device} over the last minute |
+| [ cgroup_10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cgroups.conf) | cgroup.net_packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
## Setup
diff --git a/collectors/cgroups.plugin/sys_fs_cgroup.c b/src/collectors/cgroups.plugin/sys_fs_cgroup.c
index 705edf6f7..fd30b4acc 100644
--- a/collectors/cgroups.plugin/sys_fs_cgroup.c
+++ b/src/collectors/cgroups.plugin/sys_fs_cgroup.c
@@ -85,11 +85,6 @@ struct cgroups_systemd_config_setting cgroups_systemd_options[] = {
{ .name = NULL, .setting = SYSTEMD_CGROUP_ERR },
};
-// Shared memory with information from detected cgroups
-netdata_ebpf_cgroup_shm_t shm_cgroup_ebpf = {NULL, NULL};
-int shm_fd_cgroup_ebpf = -1;
-sem_t *shm_mutex_cgroup_ebpf = SEM_FAILED;
-
struct discovery_thread discovery_thread;
@@ -255,7 +250,7 @@ void read_cgroup_plugin_configuration() {
if(cgroup_update_every < localhost->rrd_update_every)
cgroup_update_every = localhost->rrd_update_every;
- cgroup_check_for_new_every = (int)config_get_number("plugin:cgroups", "check for new cgroups every", (long long)cgroup_check_for_new_every * (long long)cgroup_update_every);
+ cgroup_check_for_new_every = (int)config_get_number("plugin:cgroups", "check for new cgroups every", cgroup_check_for_new_every);
if(cgroup_check_for_new_every < cgroup_update_every)
cgroup_check_for_new_every = cgroup_update_every;
@@ -418,6 +413,8 @@ void read_cgroup_plugin_configuration() {
" !*/init.scope " // ignore init.scope
" !/system.slice/run-*.scope " // ignore system.slice/run-XXXX.scope
+ " *user.slice/docker-*" // allow docker rootless containers
+ " !*user.slice*" // ignore the rest stuff in user.slice
" *.scope " // we need all other *.scope for sure
// ----------------------------------------------------------------
@@ -452,6 +449,7 @@ void read_cgroup_plugin_configuration() {
" !/lxc.monitor* "
" !/lxc.pivot "
" !/lxc.payload "
+ " !*lxcfs.service/.control"
" !/machine "
" !/qemu "
" !/system "
@@ -474,7 +472,6 @@ void read_cgroup_plugin_configuration() {
" !/system "
" !/systemd "
" !/user "
- " !/user.slice "
" !/lxc/*/* " // #2161 #2649
" !/lxc.monitor "
" !/lxc.payload/*/* "
@@ -526,63 +523,6 @@ void read_cgroup_plugin_configuration() {
mountinfo_free_all(root);
}
-void netdata_cgroup_ebpf_set_values(size_t length)
-{
- sem_wait(shm_mutex_cgroup_ebpf);
-
- shm_cgroup_ebpf.header->cgroup_max = cgroup_root_max;
- shm_cgroup_ebpf.header->systemd_enabled = cgroup_enable_systemd_services |
- cgroup_enable_systemd_services_detailed_memory |
- cgroup_used_memory;
- shm_cgroup_ebpf.header->body_length = length;
-
- sem_post(shm_mutex_cgroup_ebpf);
-}
-
-void netdata_cgroup_ebpf_initialize_shm()
-{
- shm_fd_cgroup_ebpf = shm_open(NETDATA_SHARED_MEMORY_EBPF_CGROUP_NAME, O_CREAT | O_RDWR, 0660);
- if (shm_fd_cgroup_ebpf < 0) {
- collector_error("Cannot initialize shared memory used by cgroup and eBPF, integration won't happen.");
- return;
- }
-
- size_t length = sizeof(netdata_ebpf_cgroup_shm_header_t) + cgroup_root_max * sizeof(netdata_ebpf_cgroup_shm_body_t);
- if (ftruncate(shm_fd_cgroup_ebpf, length)) {
- collector_error("Cannot set size for shared memory.");
- goto end_init_shm;
- }
-
- shm_cgroup_ebpf.header = (netdata_ebpf_cgroup_shm_header_t *) mmap(NULL, length,
- PROT_READ | PROT_WRITE, MAP_SHARED,
- shm_fd_cgroup_ebpf, 0);
-
- if (unlikely(MAP_FAILED == shm_cgroup_ebpf.header)) {
- shm_cgroup_ebpf.header = NULL;
- collector_error("Cannot map shared memory used between cgroup and eBPF, integration won't happen");
- goto end_init_shm;
- }
- shm_cgroup_ebpf.body = (netdata_ebpf_cgroup_shm_body_t *) ((char *)shm_cgroup_ebpf.header +
- sizeof(netdata_ebpf_cgroup_shm_header_t));
-
- shm_mutex_cgroup_ebpf = sem_open(NETDATA_NAMED_SEMAPHORE_EBPF_CGROUP_NAME, O_CREAT,
- S_IRUSR | S_IWUSR | S_IRGRP | S_IWGRP | S_IROTH | S_IWOTH, 1);
-
- if (shm_mutex_cgroup_ebpf != SEM_FAILED) {
- netdata_cgroup_ebpf_set_values(length);
- return;
- }
-
- collector_error("Cannot create semaphore, integration between eBPF and cgroup won't happen");
- munmap(shm_cgroup_ebpf.header, length);
- shm_cgroup_ebpf.header = NULL;
-
-end_init_shm:
- close(shm_fd_cgroup_ebpf);
- shm_fd_cgroup_ebpf = -1;
- shm_unlink(NETDATA_SHARED_MEMORY_EBPF_CGROUP_NAME);
-}
-
// ---------------------------------------------------------------------------------------------
static unsigned long long calc_delta(unsigned long long curr, unsigned long long prev) {
@@ -1280,12 +1220,12 @@ cpu_limits2_err:
static inline int update_memory_limits(struct cgroup *cg) {
char **filename = &cg->filename_memory_limit;
- const RRDSETVAR_ACQUIRED **chart_var = &cg->chart_var_memory_limit;
+ const RRDVAR_ACQUIRED **chart_var = &cg->chart_var_memory_limit;
unsigned long long *value = &cg->memory_limit;
if(*filename) {
if(unlikely(!*chart_var)) {
- *chart_var = rrdsetvar_custom_chart_variable_add_and_acquire(cg->st_mem_usage, "memory_limit");
+ *chart_var = rrdvar_chart_variable_add_and_acquire(cg->st_mem_usage, "memory_limit");
if(!*chart_var) {
collector_error("Cannot create cgroup %s chart variable '%s'. Will not update its limit anymore.", cg->id, "memory_limit");
freez(*filename);
@@ -1301,12 +1241,13 @@ static inline int update_memory_limits(struct cgroup *cg) {
*filename = NULL;
}
else {
- rrdsetvar_custom_chart_variable_set(cg->st_mem_usage, *chart_var, (NETDATA_DOUBLE)(*value) / (1024.0 * 1024.0));
+ rrdvar_chart_variable_set(
+ cg->st_mem_usage, *chart_var, (NETDATA_DOUBLE)(*value) / (1024.0 * 1024.0));
return 1;
}
} else {
- char buffer[30 + 1];
- int ret = read_file(*filename, buffer, 30);
+ char buffer[32];
+ int ret = read_txt_file(*filename, buffer, sizeof(buffer));
if(ret) {
collector_error("Cannot refresh cgroup %s memory limit by reading '%s'. Will not update its limit anymore.", cg->id, *filename);
freez(*filename);
@@ -1316,11 +1257,12 @@ static inline int update_memory_limits(struct cgroup *cg) {
char *s = "max\n\0";
if(strcmp(s, buffer) == 0){
*value = UINT64_MAX;
- rrdsetvar_custom_chart_variable_set(cg->st_mem_usage, *chart_var, (NETDATA_DOUBLE)(*value) / (1024.0 * 1024.0));
+ rrdvar_chart_variable_set(
+ cg->st_mem_usage, *chart_var, (NETDATA_DOUBLE)(*value) / (1024.0 * 1024.0));
return 1;
}
*value = str2ull(buffer, NULL);
- rrdsetvar_custom_chart_variable_set(cg->st_mem_usage, *chart_var, (NETDATA_DOUBLE)(*value) / (1024.0 * 1024.0));
+ rrdvar_chart_variable_set(cg->st_mem_usage, *chart_var, (NETDATA_DOUBLE)(*value) / (1024.0 * 1024.0));
return 1;
}
}
@@ -1398,7 +1340,7 @@ void update_cgroup_charts() {
}
if(unlikely(!cg->chart_var_cpu_limit)) {
- cg->chart_var_cpu_limit = rrdsetvar_custom_chart_variable_add_and_acquire(cg->st_cpu, "cpu_limit");
+ cg->chart_var_cpu_limit = rrdvar_chart_variable_add_and_acquire(cg->st_cpu, "cpu_limit");
if(!cg->chart_var_cpu_limit) {
collector_error("Cannot create cgroup %s chart variable 'cpu_limit'. Will not update its limit anymore.", cg->id);
if(cg->filename_cpuset_cpus) freez(cg->filename_cpuset_cpus);
@@ -1428,7 +1370,7 @@ void update_cgroup_charts() {
rrdset_is_obsolete___safe_from_collector_thread(cg->st_cpu_limit);
cg->st_cpu_limit = NULL;
}
- rrdsetvar_custom_chart_variable_set(cg->st_cpu, cg->chart_var_cpu_limit, NAN);
+ rrdvar_chart_variable_set(cg->st_cpu, cg->chart_var_cpu_limit, NAN);
}
}
}
@@ -1592,19 +1534,6 @@ static void cgroup_main_cleanup(void *ptr) {
}
}
- if (shm_mutex_cgroup_ebpf != SEM_FAILED) {
- sem_close(shm_mutex_cgroup_ebpf);
- }
-
- if (shm_cgroup_ebpf.header) {
- shm_cgroup_ebpf.header->cgroup_root_count = 0;
- munmap(shm_cgroup_ebpf.header, shm_cgroup_ebpf.header->body_length);
- }
-
- if (shm_fd_cgroup_ebpf > 0) {
- close(shm_fd_cgroup_ebpf);
- }
-
static_thread->enabled = NETDATA_MAIN_THREAD_EXITED;
}
@@ -1641,8 +1570,6 @@ void *cgroups_main(void *ptr) {
cgroup_read_host_total_ram();
- netdata_cgroup_ebpf_initialize_shm();
-
if (uv_mutex_init(&cgroup_root_mutex)) {
collector_error("CGROUP: cannot initialize mutex for the main cgroup list");
goto exit;
@@ -1669,10 +1596,17 @@ void *cgroups_main(void *ptr) {
// we register this only on localhost
// for the other nodes, the origin server should register it
- rrd_collector_started(); // this creates a collector that runs for as long as netdata runs
cgroup_netdev_link_init();
- rrd_function_add(localhost, NULL, "containers-vms", 10, RRDFUNCTIONS_CGTOP_HELP, true, cgroup_function_cgroup_top, NULL);
- rrd_function_add(localhost, NULL, "systemd-services", 10, RRDFUNCTIONS_CGTOP_HELP, true, cgroup_function_systemd_top, NULL);
+
+ rrd_function_add_inline(localhost, NULL, "containers-vms", 10,
+ RRDFUNCTIONS_PRIORITY_DEFAULT / 2, RRDFUNCTIONS_CGTOP_HELP,
+ "top", HTTP_ACCESS_ANONYMOUS_DATA,
+ cgroup_function_cgroup_top);
+
+ rrd_function_add_inline(localhost, NULL, "systemd-services", 10,
+ RRDFUNCTIONS_PRIORITY_DEFAULT / 3, RRDFUNCTIONS_SYSTEMD_SERVICES_HELP,
+ "top", HTTP_ACCESS_ANONYMOUS_DATA,
+ cgroup_function_systemd_top);
heartbeat_t hb;
heartbeat_init(&hb);
diff --git a/collectors/cgroups.plugin/sys_fs_cgroup.h b/src/collectors/cgroups.plugin/sys_fs_cgroup.h
index e8cfcf5f6..e8cfcf5f6 100644
--- a/collectors/cgroups.plugin/sys_fs_cgroup.h
+++ b/src/collectors/cgroups.plugin/sys_fs_cgroup.h
diff --git a/collectors/cgroups.plugin/tests/test_cgroups_plugin.c b/src/collectors/cgroups.plugin/tests/test_cgroups_plugin.c
index bb1fb3988..bb1fb3988 100644
--- a/collectors/cgroups.plugin/tests/test_cgroups_plugin.c
+++ b/src/collectors/cgroups.plugin/tests/test_cgroups_plugin.c
diff --git a/collectors/cgroups.plugin/tests/test_cgroups_plugin.h b/src/collectors/cgroups.plugin/tests/test_cgroups_plugin.h
index 3d68e9230..3d68e9230 100644
--- a/collectors/cgroups.plugin/tests/test_cgroups_plugin.h
+++ b/src/collectors/cgroups.plugin/tests/test_cgroups_plugin.h
diff --git a/collectors/cgroups.plugin/tests/test_doubles.c b/src/collectors/cgroups.plugin/tests/test_doubles.c
index b13d4b19c..53fefa9c2 100644
--- a/collectors/cgroups.plugin/tests/test_doubles.c
+++ b/src/collectors/cgroups.plugin/tests/test_doubles.c
@@ -101,7 +101,7 @@ collected_number rrddim_set_by_pointer(RRDSET *st, RRDDIM *rd, collected_number
return 0;
}
-const RRDSETVAR_ACQUIRED *rrdsetvar_custom_chart_variable_add_and_acquire(RRDSET *st, const char *name)
+const RRDVAR_ACQUIRED *rrdvar_chart_variable_add_and_acquire(RRDSET *st, const char *name)
{
UNUSED(st);
UNUSED(name);
@@ -109,7 +109,7 @@ const RRDSETVAR_ACQUIRED *rrdsetvar_custom_chart_variable_add_and_acquire(RRDSET
return NULL;
}
-void rrdsetvar_custom_chart_variable_set(RRDSET *st, const RRDSETVAR_ACQUIRED *rsa, NETDATA_DOUBLE value)
+void rrdvar_chart_variable_set(RRDSET *st, const RRDVAR_ACQUIRED *rsa, NETDATA_DOUBLE value)
{
UNUSED(st);
UNUSED(rsa);
diff --git a/collectors/charts.d.plugin/ap/README.md b/src/collectors/charts.d.plugin/ap/README.md
index 5b6e75130..5b6e75130 120000
--- a/collectors/charts.d.plugin/ap/README.md
+++ b/src/collectors/charts.d.plugin/ap/README.md
diff --git a/collectors/charts.d.plugin/ap/ap.chart.sh b/src/collectors/charts.d.plugin/ap/ap.chart.sh
index 80c9dc602..80c9dc602 100644
--- a/collectors/charts.d.plugin/ap/ap.chart.sh
+++ b/src/collectors/charts.d.plugin/ap/ap.chart.sh
diff --git a/collectors/charts.d.plugin/ap/ap.conf b/src/collectors/charts.d.plugin/ap/ap.conf
index 38fc157ce..38fc157ce 100644
--- a/collectors/charts.d.plugin/ap/ap.conf
+++ b/src/collectors/charts.d.plugin/ap/ap.conf
diff --git a/collectors/charts.d.plugin/ap/integrations/access_points.md b/src/collectors/charts.d.plugin/ap/integrations/access_points.md
index a0de2c4df..0c0a3feb6 100644
--- a/collectors/charts.d.plugin/ap/integrations/access_points.md
+++ b/src/collectors/charts.d.plugin/ap/integrations/access_points.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/ap/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/ap/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/charts.d.plugin/ap/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/charts.d.plugin/ap/metadata.yaml"
sidebar_label: "Access Points"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
+learn_rel_path: "Collecting Metrics/Linux Systems/Network"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -101,7 +101,7 @@ The configuration file name for this integration is `charts.d/ap.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/charts.d.plugin/ap/metadata.yaml b/src/collectors/charts.d.plugin/ap/metadata.yaml
index ee941e417..ee941e417 100644
--- a/collectors/charts.d.plugin/ap/metadata.yaml
+++ b/src/collectors/charts.d.plugin/ap/metadata.yaml
diff --git a/collectors/charts.d.plugin/apcupsd/README.md b/src/collectors/charts.d.plugin/apcupsd/README.md
index fc6681fe6..fc6681fe6 120000
--- a/collectors/charts.d.plugin/apcupsd/README.md
+++ b/src/collectors/charts.d.plugin/apcupsd/README.md
diff --git a/collectors/charts.d.plugin/apcupsd/apcupsd.chart.sh b/src/collectors/charts.d.plugin/apcupsd/apcupsd.chart.sh
index da9cd19c3..da9cd19c3 100644
--- a/collectors/charts.d.plugin/apcupsd/apcupsd.chart.sh
+++ b/src/collectors/charts.d.plugin/apcupsd/apcupsd.chart.sh
diff --git a/collectors/charts.d.plugin/apcupsd/apcupsd.conf b/src/collectors/charts.d.plugin/apcupsd/apcupsd.conf
index 679c0d61b..679c0d61b 100644
--- a/collectors/charts.d.plugin/apcupsd/apcupsd.conf
+++ b/src/collectors/charts.d.plugin/apcupsd/apcupsd.conf
diff --git a/collectors/charts.d.plugin/apcupsd/integrations/apc_ups.md b/src/collectors/charts.d.plugin/apcupsd/integrations/apc_ups.md
index a5c1f9613..4e3aa3352 100644
--- a/collectors/charts.d.plugin/apcupsd/integrations/apc_ups.md
+++ b/src/collectors/charts.d.plugin/apcupsd/integrations/apc_ups.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/apcupsd/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/apcupsd/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/charts.d.plugin/apcupsd/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/charts.d.plugin/apcupsd/metadata.yaml"
sidebar_label: "APC UPS"
learn_status: "Published"
-learn_rel_path: "Data Collection/UPS"
+learn_rel_path: "Collecting Metrics/UPS"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -85,16 +85,16 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ apcupsd_ups_charge ](https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf) | apcupsd.charge | average UPS charge over the last minute |
-| [ apcupsd_10min_ups_load ](https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf) | apcupsd.load | average UPS load over the last 10 minutes |
-| [ apcupsd_last_collected_secs ](https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf) | apcupsd.load | number of seconds since the last successful data collection |
-| [ apcupsd_selftest_warning ](https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf) | apcupsd.selftest | self-test failed due to insufficient battery capacity or due to overload. |
-| [ apcupsd_status_onbatt ](https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS has switched to battery power because the input power has failed |
-| [ apcupsd_status_overload ](https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS is overloaded and cannot supply enough power to the load |
-| [ apcupsd_status_lowbatt ](https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS battery is low and needs to be recharged |
-| [ apcupsd_status_replacebatt ](https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS battery has reached the end of its lifespan and needs to be replaced |
-| [ apcupsd_status_nobatt ](https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS has no battery |
-| [ apcupsd_status_commlost ](https://github.com/netdata/netdata/blob/master/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS communication link is lost |
+| [ apcupsd_ups_charge ](https://github.com/netdata/netdata/blob/master/src/health/health.d/apcupsd.conf) | apcupsd.charge | average UPS charge over the last minute |
+| [ apcupsd_10min_ups_load ](https://github.com/netdata/netdata/blob/master/src/health/health.d/apcupsd.conf) | apcupsd.load | average UPS load over the last 10 minutes |
+| [ apcupsd_last_collected_secs ](https://github.com/netdata/netdata/blob/master/src/health/health.d/apcupsd.conf) | apcupsd.load | number of seconds since the last successful data collection |
+| [ apcupsd_selftest_warning ](https://github.com/netdata/netdata/blob/master/src/health/health.d/apcupsd.conf) | apcupsd.selftest | self-test failed due to insufficient battery capacity or due to overload. |
+| [ apcupsd_status_onbatt ](https://github.com/netdata/netdata/blob/master/src/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS has switched to battery power because the input power has failed |
+| [ apcupsd_status_overload ](https://github.com/netdata/netdata/blob/master/src/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS is overloaded and cannot supply enough power to the load |
+| [ apcupsd_status_lowbatt ](https://github.com/netdata/netdata/blob/master/src/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS battery is low and needs to be recharged |
+| [ apcupsd_status_replacebatt ](https://github.com/netdata/netdata/blob/master/src/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS battery has reached the end of its lifespan and needs to be replaced |
+| [ apcupsd_status_nobatt ](https://github.com/netdata/netdata/blob/master/src/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS has no battery |
+| [ apcupsd_status_commlost ](https://github.com/netdata/netdata/blob/master/src/health/health.d/apcupsd.conf) | apcupsd.status | APC UPS communication link is lost |
## Setup
@@ -119,7 +119,7 @@ The configuration file name for this integration is `charts.d/apcupsd.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/charts.d.plugin/charts.d.conf b/src/collectors/charts.d.plugin/charts.d.conf
index 4614f259e..4614f259e 100644
--- a/collectors/charts.d.plugin/charts.d.conf
+++ b/src/collectors/charts.d.plugin/charts.d.conf
diff --git a/collectors/charts.d.plugin/charts.d.dryrun-helper.sh b/src/collectors/charts.d.plugin/charts.d.dryrun-helper.sh
index 91af2c542..91af2c542 100755
--- a/collectors/charts.d.plugin/charts.d.dryrun-helper.sh
+++ b/src/collectors/charts.d.plugin/charts.d.dryrun-helper.sh
diff --git a/collectors/charts.d.plugin/charts.d.plugin.in b/src/collectors/charts.d.plugin/charts.d.plugin.in
index 4e64b7e23..4e64b7e23 100755
--- a/collectors/charts.d.plugin/charts.d.plugin.in
+++ b/src/collectors/charts.d.plugin/charts.d.plugin.in
diff --git a/collectors/charts.d.plugin/example/example.chart.sh b/src/collectors/charts.d.plugin/example/example.chart.sh
index 6bbbcf1d7..6bbbcf1d7 100644
--- a/collectors/charts.d.plugin/example/example.chart.sh
+++ b/src/collectors/charts.d.plugin/example/example.chart.sh
diff --git a/collectors/charts.d.plugin/example/example.conf b/src/collectors/charts.d.plugin/example/example.conf
index 6232ca584..6232ca584 100644
--- a/collectors/charts.d.plugin/example/example.conf
+++ b/src/collectors/charts.d.plugin/example/example.conf
diff --git a/collectors/charts.d.plugin/libreswan/README.md b/src/collectors/charts.d.plugin/libreswan/README.md
index 1416d9597..1416d9597 120000
--- a/collectors/charts.d.plugin/libreswan/README.md
+++ b/src/collectors/charts.d.plugin/libreswan/README.md
diff --git a/collectors/charts.d.plugin/libreswan/integrations/libreswan.md b/src/collectors/charts.d.plugin/libreswan/integrations/libreswan.md
index bd1eec647..0b08e2b4f 100644
--- a/collectors/charts.d.plugin/libreswan/integrations/libreswan.md
+++ b/src/collectors/charts.d.plugin/libreswan/integrations/libreswan.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/libreswan/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/libreswan/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/charts.d.plugin/libreswan/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/charts.d.plugin/libreswan/metadata.yaml"
sidebar_label: "Libreswan"
learn_status: "Published"
-learn_rel_path: "Data Collection/VPNs"
+learn_rel_path: "Collecting Metrics/VPNs"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -116,7 +116,7 @@ The configuration file name for this integration is `charts.d/libreswan.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/charts.d.plugin/libreswan/libreswan.chart.sh b/src/collectors/charts.d.plugin/libreswan/libreswan.chart.sh
index d526f7a91..d526f7a91 100644
--- a/collectors/charts.d.plugin/libreswan/libreswan.chart.sh
+++ b/src/collectors/charts.d.plugin/libreswan/libreswan.chart.sh
diff --git a/collectors/charts.d.plugin/libreswan/libreswan.conf b/src/collectors/charts.d.plugin/libreswan/libreswan.conf
index 9b3ee77b7..9b3ee77b7 100644
--- a/collectors/charts.d.plugin/libreswan/libreswan.conf
+++ b/src/collectors/charts.d.plugin/libreswan/libreswan.conf
diff --git a/collectors/charts.d.plugin/libreswan/metadata.yaml b/src/collectors/charts.d.plugin/libreswan/metadata.yaml
index 77cb25450..77cb25450 100644
--- a/collectors/charts.d.plugin/libreswan/metadata.yaml
+++ b/src/collectors/charts.d.plugin/libreswan/metadata.yaml
diff --git a/collectors/charts.d.plugin/loopsleepms.sh.inc b/src/collectors/charts.d.plugin/loopsleepms.sh.inc
index 5608b8d8f..5608b8d8f 100644
--- a/collectors/charts.d.plugin/loopsleepms.sh.inc
+++ b/src/collectors/charts.d.plugin/loopsleepms.sh.inc
diff --git a/collectors/charts.d.plugin/opensips/README.md b/src/collectors/charts.d.plugin/opensips/README.md
index bb85ba6d0..bb85ba6d0 120000
--- a/collectors/charts.d.plugin/opensips/README.md
+++ b/src/collectors/charts.d.plugin/opensips/README.md
diff --git a/collectors/charts.d.plugin/opensips/integrations/opensips.md b/src/collectors/charts.d.plugin/opensips/integrations/opensips.md
index 8c88dba0b..6a6a309e8 100644
--- a/collectors/charts.d.plugin/opensips/integrations/opensips.md
+++ b/src/collectors/charts.d.plugin/opensips/integrations/opensips.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/opensips/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/opensips/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/charts.d.plugin/opensips/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/charts.d.plugin/opensips/metadata.yaml"
sidebar_label: "OpenSIPS"
learn_status: "Published"
-learn_rel_path: "Data Collection/Telephony Servers"
+learn_rel_path: "Collecting Metrics/Telephony Servers"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -112,7 +112,7 @@ The configuration file name for this integration is `charts.d/opensips.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/charts.d.plugin/opensips/metadata.yaml b/src/collectors/charts.d.plugin/opensips/metadata.yaml
index 356de5615..356de5615 100644
--- a/collectors/charts.d.plugin/opensips/metadata.yaml
+++ b/src/collectors/charts.d.plugin/opensips/metadata.yaml
diff --git a/collectors/charts.d.plugin/opensips/opensips.chart.sh b/src/collectors/charts.d.plugin/opensips/opensips.chart.sh
index 02401fd59..02401fd59 100644
--- a/collectors/charts.d.plugin/opensips/opensips.chart.sh
+++ b/src/collectors/charts.d.plugin/opensips/opensips.chart.sh
diff --git a/collectors/charts.d.plugin/opensips/opensips.conf b/src/collectors/charts.d.plugin/opensips/opensips.conf
index e25111dce..e25111dce 100644
--- a/collectors/charts.d.plugin/opensips/opensips.conf
+++ b/src/collectors/charts.d.plugin/opensips/opensips.conf
diff --git a/collectors/charts.d.plugin/sensors/README.md b/src/collectors/charts.d.plugin/sensors/README.md
index 7e5a416c4..7e5a416c4 120000
--- a/collectors/charts.d.plugin/sensors/README.md
+++ b/src/collectors/charts.d.plugin/sensors/README.md
diff --git a/collectors/charts.d.plugin/sensors/integrations/linux_sensors_sysfs.md b/src/collectors/charts.d.plugin/sensors/integrations/linux_sensors_sysfs.md
index 130352f61..6e802bb7e 100644
--- a/collectors/charts.d.plugin/sensors/integrations/linux_sensors_sysfs.md
+++ b/src/collectors/charts.d.plugin/sensors/integrations/linux_sensors_sysfs.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/sensors/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/charts.d.plugin/sensors/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/charts.d.plugin/sensors/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/charts.d.plugin/sensors/metadata.yaml"
sidebar_label: "Linux Sensors (sysfs)"
learn_status: "Published"
-learn_rel_path: "Data Collection/Hardware Devices and Sensors"
+learn_rel_path: "Collecting Metrics/Hardware Devices and Sensors"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -22,7 +22,7 @@ Module: sensors
## Overview
Use this collector when `lm-sensors` doesn't work on your device (e.g. for RPi temperatures).
-For all other cases use the [Python collector](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/sensors), which supports multiple jobs, is more efficient and performs calculations on top of the kernel provided values."
+For all other cases use the [Python collector](https://github.com/netdata/netdata/blob/master/src/collectors/python.d.plugin/sensors), which supports multiple jobs, is more efficient and performs calculations on top of the kernel provided values."
It will provide charts for all configured system sensors, by reading sensors directly from the kernel.
@@ -95,14 +95,14 @@ If [using our official native DEB/RPM packages](https://github.com/netdata/netda
#### Enable the sensors collector
-The `sensors` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`, to edit the `charts.d.conf` file.
+The `sensors` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md), which is typically at `/etc/netdata`, to edit the `charts.d.conf` file.
```bash
cd /etc/netdata # Replace this path with your Netdata config directory, if different
sudo ./edit-config charts.d.conf
```
-Change the value of the `sensors` setting to `force` and uncomment the line. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
+Change the value of the `sensors` setting to `force` and uncomment the line. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/packaging/installer/README.md#maintaining-a-netdata-agent-installation) for your system.
@@ -114,7 +114,7 @@ The configuration file name for this integration is `charts.d/sensors.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/charts.d.plugin/sensors/sensors.chart.sh b/src/collectors/charts.d.plugin/sensors/sensors.chart.sh
index 9576e2ab2..9576e2ab2 100644
--- a/collectors/charts.d.plugin/sensors/sensors.chart.sh
+++ b/src/collectors/charts.d.plugin/sensors/sensors.chart.sh
diff --git a/collectors/charts.d.plugin/sensors/sensors.conf b/src/collectors/charts.d.plugin/sensors/sensors.conf
index bcb28807d..bcb28807d 100644
--- a/collectors/charts.d.plugin/sensors/sensors.conf
+++ b/src/collectors/charts.d.plugin/sensors/sensors.conf
diff --git a/collectors/cups.plugin/README.md b/src/collectors/cups.plugin/README.md
index e32570639..e32570639 120000
--- a/collectors/cups.plugin/README.md
+++ b/src/collectors/cups.plugin/README.md
diff --git a/collectors/cups.plugin/cups_plugin.c b/src/collectors/cups.plugin/cups_plugin.c
index 8efd32e31..710e8f87d 100644
--- a/collectors/cups.plugin/cups_plugin.c
+++ b/src/collectors/cups.plugin/cups_plugin.c
@@ -175,7 +175,7 @@ struct job_metrics *get_job_metrics(char *dest) {
struct job_metrics new_job_metrics = { .id = ++job_id };
jm = dictionary_set(dict_dest_job_metrics, dest, &new_job_metrics, sizeof(struct job_metrics));
send_job_charts_definitions_to_netdata(dest, jm->id, false);
- };
+ }
return jm;
}
diff --git a/collectors/cups.plugin/integrations/cups.md b/src/collectors/cups.plugin/integrations/cups.md
index a8ea5b15f..6aacd6b81 100644
--- a/collectors/cups.plugin/integrations/cups.md
+++ b/src/collectors/cups.plugin/integrations/cups.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/cups.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/cups.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/cups.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/cups.plugin/metadata.yaml"
sidebar_label: "CUPS"
learn_status: "Published"
-learn_rel_path: "Data Collection/Hardware Devices and Sensors"
+learn_rel_path: "Collecting Metrics/Hardware Devices and Sensors"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -116,7 +116,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/cups.plugin/metadata.yaml b/src/collectors/cups.plugin/metadata.yaml
index 9ec2f4118..9ec2f4118 100644
--- a/collectors/cups.plugin/metadata.yaml
+++ b/src/collectors/cups.plugin/metadata.yaml
diff --git a/collectors/debugfs.plugin/debugfs_extfrag.c b/src/collectors/debugfs.plugin/debugfs_extfrag.c
index 75da4deca..75da4deca 100644
--- a/collectors/debugfs.plugin/debugfs_extfrag.c
+++ b/src/collectors/debugfs.plugin/debugfs_extfrag.c
diff --git a/collectors/debugfs.plugin/debugfs_plugin.c b/src/collectors/debugfs.plugin/debugfs_plugin.c
index 13012ec40..94e3db631 100644
--- a/collectors/debugfs.plugin/debugfs_plugin.c
+++ b/src/collectors/debugfs.plugin/debugfs_plugin.c
@@ -28,7 +28,7 @@ static struct debugfs_module {
{ .name = NULL, .enabled = CONFIG_BOOLEAN_NO, .func = NULL}
};
-#ifdef HAVE_CAPABILITY
+#ifdef HAVE_SYS_CAPABILITY_H
static int debugfs_check_capabilities()
{
cap_t caps = cap_get_proc();
@@ -182,7 +182,7 @@ int main(int argc, char **argv)
// FIXME: remove debugfs_check_sys_permission() after https://github.com/netdata/netdata/issues/15048 is fixed
if (!debugfs_check_capabilities() && !debugfs_am_i_running_as_root() && !debugfs_check_sys_permission()) {
uid_t uid = getuid(), euid = geteuid();
-#ifdef HAVE_CAPABILITY
+#ifdef HAVE_SYS_CAPABILITY_H
netdata_log_error(
"debugfs.plugin should either run as root (now running with uid %u, euid %u) or have special capabilities. "
"Without these, debugfs.plugin cannot access /sys/kernel/debug. "
diff --git a/collectors/debugfs.plugin/debugfs_plugin.h b/src/collectors/debugfs.plugin/debugfs_plugin.h
index 903e4a19e..903e4a19e 100644
--- a/collectors/debugfs.plugin/debugfs_plugin.h
+++ b/src/collectors/debugfs.plugin/debugfs_plugin.h
diff --git a/collectors/debugfs.plugin/debugfs_zswap.c b/src/collectors/debugfs.plugin/debugfs_zswap.c
index 502a04f1f..e4d956cda 100644
--- a/collectors/debugfs.plugin/debugfs_zswap.c
+++ b/src/collectors/debugfs.plugin/debugfs_zswap.c
@@ -370,7 +370,7 @@ static int debugfs_is_zswap_enabled()
snprintfz(filename, FILENAME_MAX, "/sys/module/zswap/parameters/enabled"); // host prefix is not needed here
char state[ZSWAP_STATE_SIZE + 1];
- int ret = read_file(filename, state, ZSWAP_STATE_SIZE);
+ int ret = read_txt_file(filename, state, sizeof(state));
if (unlikely(!ret && !strcmp(state, "Y"))) {
return 0;
diff --git a/collectors/debugfs.plugin/integrations/linux_zswap.md b/src/collectors/debugfs.plugin/integrations/linux_zswap.md
index 44478454b..20de4c78d 100644
--- a/collectors/debugfs.plugin/integrations/linux_zswap.md
+++ b/src/collectors/debugfs.plugin/integrations/linux_zswap.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/debugfs.plugin/integrations/linux_zswap.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/debugfs.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/debugfs.plugin/integrations/linux_zswap.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/debugfs.plugin/metadata.yaml"
sidebar_label: "Linux ZSwap"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Memory"
+learn_rel_path: "Collecting Metrics/Linux Systems/Memory"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -113,7 +113,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/debugfs.plugin/integrations/power_capping.md b/src/collectors/debugfs.plugin/integrations/power_capping.md
index d4b7eb890..de68ff542 100644
--- a/collectors/debugfs.plugin/integrations/power_capping.md
+++ b/src/collectors/debugfs.plugin/integrations/power_capping.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/debugfs.plugin/integrations/power_capping.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/debugfs.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/debugfs.plugin/integrations/power_capping.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/debugfs.plugin/metadata.yaml"
sidebar_label: "Power Capping"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Kernel"
+learn_rel_path: "Collecting Metrics/Linux Systems/Kernel"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -107,7 +107,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/debugfs.plugin/integrations/system_memory_fragmentation.md b/src/collectors/debugfs.plugin/integrations/system_memory_fragmentation.md
index ef287bc30..8c5f7317a 100644
--- a/collectors/debugfs.plugin/integrations/system_memory_fragmentation.md
+++ b/src/collectors/debugfs.plugin/integrations/system_memory_fragmentation.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/debugfs.plugin/integrations/system_memory_fragmentation.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/debugfs.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/debugfs.plugin/integrations/system_memory_fragmentation.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/debugfs.plugin/metadata.yaml"
sidebar_label: "System Memory Fragmentation"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Memory"
+learn_rel_path: "Collecting Metrics/Linux Systems/Memory"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -111,7 +111,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/debugfs.plugin/metadata.yaml b/src/collectors/debugfs.plugin/metadata.yaml
index d3bf0a0d8..d3bf0a0d8 100644
--- a/collectors/debugfs.plugin/metadata.yaml
+++ b/src/collectors/debugfs.plugin/metadata.yaml
diff --git a/collectors/debugfs.plugin/sys_devices_virtual_powercap.c b/src/collectors/debugfs.plugin/sys_devices_virtual_powercap.c
index ee261c27f..a5dfb7550 100644
--- a/collectors/debugfs.plugin/sys_devices_virtual_powercap.c
+++ b/src/collectors/debugfs.plugin/sys_devices_virtual_powercap.c
@@ -27,7 +27,7 @@ static struct zone_t *get_rapl_zone(const char *control_type __maybe_unused, str
snprintfz(temp, FILENAME_MAX, "%s/%s", dirname, "name");
char name[FILENAME_MAX + 1] = "";
- if (read_file(temp, name, sizeof(name) - 1) != 0)
+ if (read_txt_file(temp, name, sizeof(name)) != 0)
return NULL;
char *trimmed = trim(name);
diff --git a/collectors/diskspace.plugin/README.md b/src/collectors/diskspace.plugin/README.md
index c9f4e1c5e..c9f4e1c5e 120000
--- a/collectors/diskspace.plugin/README.md
+++ b/src/collectors/diskspace.plugin/README.md
diff --git a/collectors/diskspace.plugin/integrations/disk_space.md b/src/collectors/diskspace.plugin/integrations/disk_space.md
index 1c937ed7f..47d3a9b67 100644
--- a/collectors/diskspace.plugin/integrations/disk_space.md
+++ b/src/collectors/diskspace.plugin/integrations/disk_space.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/diskspace.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/diskspace.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/diskspace.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/diskspace.plugin/metadata.yaml"
sidebar_label: "Disk space"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems"
+learn_rel_path: "Collecting Metrics/Linux Systems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -81,8 +81,8 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ disk_space_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf) | disk.space | disk ${label:mount_point} space utilization |
-| [ disk_inode_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf) | disk.inodes | disk ${label:mount_point} inode utilization |
+| [ disk_space_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/disks.conf) | disk.space | disk ${label:mount_point} space utilization |
+| [ disk_inode_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/disks.conf) | disk.inodes | disk ${label:mount_point} inode utilization |
## Setup
@@ -109,7 +109,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/diskspace.plugin/plugin_diskspace.c b/src/collectors/diskspace.plugin/plugin_diskspace.c
index 94257810c..d93e743ae 100644
--- a/collectors/diskspace.plugin/plugin_diskspace.c
+++ b/src/collectors/diskspace.plugin/plugin_diskspace.c
@@ -40,11 +40,10 @@ static inline void mountinfo_reload(int force) {
struct mount_point_metadata {
int do_space;
int do_inodes;
- int shown_error;
- int updated;
- int slow;
- bool function_ready;
+ bool shown_error;
+ bool updated;
+ bool slow;
STRING *filesystem;
STRING *mountroot;
@@ -68,48 +67,44 @@ static DICTIONARY *dict_mountpoints = NULL;
#define rrdset_obsolete_and_pointer_null(st) do { if(st) { rrdset_is_obsolete___safe_from_collector_thread(st); (st) = NULL; } } while(st)
-int mount_point_cleanup(const char *name, void *entry, int slow) {
- (void)name;
-
- struct mount_point_metadata *mp = (struct mount_point_metadata *)entry;
- if(!mp) return 0;
-
- if (slow != mp->slow)
- return 0;
+static void mount_points_cleanup(bool slow) {
+ struct mount_point_metadata *mp;
+ dfe_start_write(dict_mountpoints, mp) {
+ if(mp->slow != slow) continue;
- if(likely(mp->updated)) {
- mp->updated = 0;
- return 0;
+ if(mp->updated)
+ mp->updated = false;
+ else if(cleanup_mount_points)
+ dictionary_del(dict_mountpoints, mp_dfe.name);
}
+ dfe_done(mp);
- if(likely(cleanup_mount_points && mp->collected)) {
- mp->function_ready = false;
- mp->collected = 0;
- mp->updated = 0;
- mp->shown_error = 0;
+ dictionary_garbage_collect(dict_mountpoints);
+}
- string_freez(mp->filesystem);
- string_freez(mp->mountroot);
+void mountpoint_delete_cb(const DICTIONARY_ITEM *item __maybe_unused, void *entry, void *data __maybe_unused) {
+ struct mount_point_metadata *mp = (struct mount_point_metadata *)entry;
- rrdset_obsolete_and_pointer_null(mp->st_space);
- rrdset_obsolete_and_pointer_null(mp->st_inodes);
+ mp->collected = 0;
+ mp->updated = false;
+ mp->shown_error = false;
- mp->rd_space_avail = NULL;
- mp->rd_space_used = NULL;
- mp->rd_space_reserved = NULL;
+ string_freez(mp->filesystem);
+ mp->filesystem = NULL;
- mp->rd_inodes_avail = NULL;
- mp->rd_inodes_used = NULL;
- mp->rd_inodes_reserved = NULL;
- }
+ string_freez(mp->mountroot);
+ mp->mountroot = NULL;
- return 0;
-}
+ rrdset_obsolete_and_pointer_null(mp->st_space);
+ rrdset_obsolete_and_pointer_null(mp->st_inodes);
-int mount_point_cleanup_cb(const DICTIONARY_ITEM *item, void *entry, void *data __maybe_unused) {
- const char *name = dictionary_acquired_item_name(item);
+ mp->rd_space_avail = NULL;
+ mp->rd_space_used = NULL;
+ mp->rd_space_reserved = NULL;
- return mount_point_cleanup(name, (struct mount_point_metadata *)entry, 0);
+ mp->rd_inodes_avail = NULL;
+ mp->rd_inodes_used = NULL;
+ mp->rd_inodes_reserved = NULL;
}
// a copy of basic mountinfo fields
@@ -148,7 +143,7 @@ static void add_basic_mountinfo(struct basic_mountinfo **root, struct mountinfo
bmi->next = *root;
*root = bmi;
-};
+}
static void free_basic_mountinfo(struct basic_mountinfo *bmi)
{
@@ -160,7 +155,7 @@ static void free_basic_mountinfo(struct basic_mountinfo *bmi)
freez(bmi);
}
-};
+}
static void free_basic_mountinfo_list(struct basic_mountinfo *root)
{
@@ -297,8 +292,6 @@ static void calculate_values_and_show_charts(
rendered++;
}
- m->function_ready = rendered > 0;
-
if(likely(rendered))
m->collected++;
}
@@ -341,11 +334,12 @@ static inline void do_disk_space_stats(struct mountinfo *mi, int update_every) {
true);
dict_mountpoints = dictionary_create_advanced(DICT_OPTION_NONE, &dictionary_stats_category_collectors, 0);
+ dictionary_register_delete_callback(dict_mountpoints, mountpoint_delete_cb, NULL);
}
- struct mount_point_metadata *m = dictionary_get(dict_mountpoints, mi->mount_point);
- if(unlikely(!m)) {
- int slow = 0;
+ const DICTIONARY_ITEM *item = dictionary_get_and_acquire_item(dict_mountpoints, mi->mount_point);
+ if(unlikely(!item)) {
+ bool slow = false;
int def_space = config_get_boolean_ondemand(CONFIG_SECTION_DISKSPACE, "space usage for all disks", CONFIG_BOOLEAN_AUTO);
int def_inodes = config_get_boolean_ondemand(CONFIG_SECTION_DISKSPACE, "inodes usage for all disks", CONFIG_BOOLEAN_AUTO);
@@ -393,7 +387,7 @@ static inline void do_disk_space_stats(struct mountinfo *mi, int update_every) {
}
if ((now_monotonic_high_precision_usec() - start_time) > slow_timeout)
- slow = 1;
+ slow = true;
}
char var_name[4096 + 1];
@@ -408,55 +402,55 @@ static inline void do_disk_space_stats(struct mountinfo *mi, int update_every) {
do_inodes = config_get_boolean_ondemand(var_name, "inodes usage", def_inodes);
struct mount_point_metadata mp = {
- .do_space = do_space,
- .do_inodes = do_inodes,
- .shown_error = 0,
- .updated = 0,
- .slow = 0,
-
- .collected = 0,
-
- .st_space = NULL,
- .rd_space_avail = NULL,
- .rd_space_used = NULL,
- .rd_space_reserved = NULL,
-
- .st_inodes = NULL,
- .rd_inodes_avail = NULL,
- .rd_inodes_used = NULL,
- .rd_inodes_reserved = NULL
+ .do_space = do_space,
+ .do_inodes = do_inodes,
+ .shown_error = false,
+ .updated = false,
+ .slow = slow,
+
+ .collected = 0,
+ .filesystem = string_strdupz(mi->filesystem),
+ .mountroot = string_strdupz(mi->root),
+ .chart_labels = rrdlabels_create(),
+
+ .st_space = NULL,
+ .rd_space_avail = NULL,
+ .rd_space_used = NULL,
+ .rd_space_reserved = NULL,
+
+ .st_inodes = NULL,
+ .rd_inodes_avail = NULL,
+ .rd_inodes_used = NULL,
+ .rd_inodes_reserved = NULL
};
- mp.filesystem = string_strdupz(mi->filesystem);
- mp.mountroot = string_strdupz(mi->root);
-
- mp.chart_labels = rrdlabels_create();
rrdlabels_add(mp.chart_labels, "mount_point", mi->mount_point, RRDLABEL_SRC_AUTO);
rrdlabels_add(mp.chart_labels, "filesystem", mi->filesystem, RRDLABEL_SRC_AUTO);
rrdlabels_add(mp.chart_labels, "mount_root", mi->root, RRDLABEL_SRC_AUTO);
- m = dictionary_set(dict_mountpoints, mi->mount_point, &mp, sizeof(struct mount_point_metadata));
-
- m->slow = slow;
+ item = dictionary_set_and_acquire_item(dict_mountpoints, mi->mount_point, &mp, sizeof(struct mount_point_metadata));
}
+ struct mount_point_metadata *m = dictionary_acquired_item_value(item);
if (m->slow) {
add_basic_mountinfo(&slow_mountinfo_tmp_root, mi);
- return;
+ goto cleanup;
}
- m->updated = 1;
+ m->updated = true;
- if(unlikely(m->do_space == CONFIG_BOOLEAN_NO && m->do_inodes == CONFIG_BOOLEAN_NO))
- return;
+ if(unlikely(m->do_space == CONFIG_BOOLEAN_NO && m->do_inodes == CONFIG_BOOLEAN_NO)) {
+ goto cleanup;
+ }
if (unlikely(
mi->flags & MOUNTINFO_READONLY &&
!(mi->flags & MOUNTINFO_IS_IN_SYSD_PROTECTED_LIST) &&
!m->collected &&
m->do_space != CONFIG_BOOLEAN_YES &&
- m->do_inodes != CONFIG_BOOLEAN_YES))
- return;
+ m->do_inodes != CONFIG_BOOLEAN_YES)) {
+ goto cleanup;
+ }
usec_t start_time = now_monotonic_high_precision_usec();
struct statvfs buff_statvfs;
@@ -469,15 +463,15 @@ static inline void do_disk_space_stats(struct mountinfo *mi, int update_every) {
, mi->filesystem?mi->filesystem:""
, mi->root?mi->root:""
);
- m->shown_error = 1;
+ m->shown_error = true;
}
- return;
+ goto cleanup;
}
if ((now_monotonic_high_precision_usec() - start_time) > slow_timeout)
- m->slow = 1;
+ m->slow = true;
- m->shown_error = 0;
+ m->shown_error = false;
struct basic_mountinfo bmi;
bmi.mount_point = mi->mount_point;
@@ -486,12 +480,17 @@ static inline void do_disk_space_stats(struct mountinfo *mi, int update_every) {
bmi.root = mi->root;
calculate_values_and_show_charts(&bmi, m, &buff_statvfs, update_every);
+
+cleanup:
+ dictionary_acquired_item_release(dict_mountpoints, item);
}
static inline void do_slow_disk_space_stats(struct basic_mountinfo *mi, int update_every) {
- struct mount_point_metadata *m = dictionary_get(dict_mountpoints, mi->mount_point);
+ const DICTIONARY_ITEM *item = dictionary_get_and_acquire_item(dict_mountpoints, mi->mount_point);
+ if(!item) return;
- m->updated = 1;
+ struct mount_point_metadata *m = dictionary_acquired_item_value(item);
+ m->updated = true;
struct statvfs buff_statvfs;
if (statvfs(mi->mount_point, &buff_statvfs) < 0) {
@@ -502,13 +501,16 @@ static inline void do_slow_disk_space_stats(struct basic_mountinfo *mi, int upda
, mi->filesystem?mi->filesystem:""
, mi->root?mi->root:""
);
- m->shown_error = 1;
+ m->shown_error = true;
}
- return;
+ goto cleanup;
}
- m->shown_error = 0;
+ m->shown_error = false;
calculate_values_and_show_charts(mi, m, &buff_statvfs, update_every);
+
+cleanup:
+ dictionary_acquired_item_release(dict_mountpoints, item);
}
static void diskspace_slow_worker_cleanup(void *ptr)
@@ -584,13 +586,9 @@ void *diskspace_slow_worker(void *ptr)
if(unlikely(!service_running(SERVICE_COLLECTORS))) break;
- worker_is_busy(WORKER_JOB_SLOW_CLEANUP);
-
- for(bmi = slow_mountinfo_root; bmi; bmi = bmi->next) {
- struct mount_point_metadata *m = dictionary_get(dict_mountpoints, bmi->mount_point);
-
- if (m)
- mount_point_cleanup(bmi->mount_point, m, 1);
+ if(dict_mountpoints) {
+ worker_is_busy(WORKER_JOB_SLOW_CLEANUP);
+ mount_points_cleanup(true);
}
usec_t dt = now_monotonic_high_precision_usec() - start_time;
@@ -636,12 +634,8 @@ static void diskspace_main_cleanup(void *ptr) {
#error WORKER_UTILIZATION_MAX_JOB_TYPES has to be at least 3
#endif
-int diskspace_function_mount_points(BUFFER *wb, int timeout __maybe_unused, const char *function __maybe_unused,
- void *collector_data __maybe_unused,
- rrd_function_result_callback_t result_cb, void *result_cb_data,
- rrd_function_is_cancelled_cb_t is_cancelled_cb, void *is_cancelled_cb_data,
- rrd_function_register_canceller_cb_t register_canceller_cb __maybe_unused,
- void *register_canceller_cb_data __maybe_unused) {
+int diskspace_function_mount_points(BUFFER *wb, const char *function __maybe_unused) {
+ netdata_mutex_lock(&slow_mountinfo_mutex);
buffer_flush(wb);
wb->content_type = CT_APPLICATION_JSON;
@@ -651,6 +645,7 @@ int diskspace_function_mount_points(BUFFER *wb, int timeout __maybe_unused, cons
buffer_json_member_add_uint64(wb, "status", HTTP_RESP_OK);
buffer_json_member_add_string(wb, "type", "table");
buffer_json_member_add_time_t(wb, "update_every", 1);
+ buffer_json_member_add_boolean(wb, "has_history", false);
buffer_json_member_add_string(wb, "help", RRDFUNCTIONS_DISKSPACE_HELP);
buffer_json_member_add_array(wb, "data");
@@ -665,8 +660,8 @@ int diskspace_function_mount_points(BUFFER *wb, int timeout __maybe_unused, cons
double max_inodes_reserved = 0.0;
struct mount_point_metadata *mp;
- dfe_start_write(dict_mountpoints, mp) {
- if (!mp->function_ready)
+ dfe_start_read(dict_mountpoints, mp) {
+ if (!mp->collected)
continue;
buffer_json_add_array_item_array(wb);
@@ -846,16 +841,8 @@ int diskspace_function_mount_points(BUFFER *wb, int timeout __maybe_unused, cons
buffer_json_member_add_time_t(wb, "expires", now_realtime_sec() + 1);
buffer_json_finalize(wb);
- int response = HTTP_RESP_OK;
- if(is_cancelled_cb && is_cancelled_cb(is_cancelled_cb_data)) {
- buffer_flush(wb);
- response = HTTP_RESP_CLIENT_CLOSED_REQUEST;
- }
-
- if(result_cb)
- result_cb(wb, response, result_cb_data);
-
- return response;
+ netdata_mutex_unlock(&slow_mountinfo_mutex);
+ return HTTP_RESP_OK;
}
void *diskspace_main(void *ptr) {
@@ -864,8 +851,10 @@ void *diskspace_main(void *ptr) {
worker_register_job_name(WORKER_JOB_MOUNTPOINT, "mountpoint");
worker_register_job_name(WORKER_JOB_CLEANUP, "cleanup");
- rrd_collector_started();
- rrd_function_add(localhost, NULL, "mount-points", 10, RRDFUNCTIONS_DISKSPACE_HELP, true, diskspace_function_mount_points, NULL);
+ rrd_function_add_inline(localhost, NULL, "mount-points", 10,
+ RRDFUNCTIONS_PRIORITY_DEFAULT, RRDFUNCTIONS_DISKSPACE_HELP,
+ "top", HTTP_ACCESS_ANONYMOUS_DATA,
+ diskspace_function_mount_points);
netdata_thread_cleanup_push(diskspace_main_cleanup, ptr);
@@ -934,9 +923,8 @@ void *diskspace_main(void *ptr) {
if(dict_mountpoints) {
worker_is_busy(WORKER_JOB_CLEANUP);
- dictionary_walkthrough_read(dict_mountpoints, mount_point_cleanup_cb, NULL);
+ mount_points_cleanup(false);
}
-
}
worker_unregister();
diff --git a/collectors/ebpf.plugin/ebpf.d.conf b/src/collectors/ebpf.plugin/ebpf.d.conf
index 5cb844b20..833c8fd99 100644
--- a/collectors/ebpf.plugin/ebpf.d.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d.conf
@@ -58,20 +58,20 @@
# When plugin detects that system has support to BTF, it enables integration with apps.plugin.
#
[ebpf programs]
- cachestat = yes
+ cachestat = no
dcstat = no
disk = no
- fd = yes
+ fd = no
filesystem = no
hardirq = no
mdflush = no
mount = yes
oomkill = yes
- process = yes
- shm = yes
+ process = no
+ shm = no
socket = no
softirq = yes
sync = no
- swap = yes
+ swap = no
vfs = no
network connections = no
diff --git a/collectors/ebpf.plugin/ebpf.d/cachestat.conf b/src/collectors/ebpf.plugin/ebpf.d/cachestat.conf
index 9c51b2c52..c378e82e8 100644
--- a/collectors/ebpf.plugin/ebpf.d/cachestat.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/cachestat.conf
@@ -37,6 +37,6 @@
# pid table size = 32768
ebpf type format = auto
ebpf co-re tracing = trampoline
- collect pid = real parent
+ collect pid = all
# maps per core = yes
lifetime = 300
diff --git a/collectors/ebpf.plugin/ebpf.d/dcstat.conf b/src/collectors/ebpf.plugin/ebpf.d/dcstat.conf
index 614d814e6..2d54bce97 100644
--- a/collectors/ebpf.plugin/ebpf.d/dcstat.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/dcstat.conf
@@ -35,6 +35,6 @@
# pid table size = 32768
ebpf type format = auto
ebpf co-re tracing = trampoline
- collect pid = real parent
+ collect pid = all
# maps per core = yes
lifetime = 300
diff --git a/collectors/ebpf.plugin/ebpf.d/disk.conf b/src/collectors/ebpf.plugin/ebpf.d/disk.conf
index c5a0a2708..c5a0a2708 100644
--- a/collectors/ebpf.plugin/ebpf.d/disk.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/disk.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/ebpf_kernel_reject_list.txt b/src/collectors/ebpf.plugin/ebpf.d/ebpf_kernel_reject_list.txt
index 539bf357f..539bf357f 100644
--- a/collectors/ebpf.plugin/ebpf.d/ebpf_kernel_reject_list.txt
+++ b/src/collectors/ebpf.plugin/ebpf.d/ebpf_kernel_reject_list.txt
diff --git a/collectors/ebpf.plugin/ebpf.d/fd.conf b/src/collectors/ebpf.plugin/ebpf.d/fd.conf
index d48230323..d48230323 100644
--- a/collectors/ebpf.plugin/ebpf.d/fd.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/fd.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/filesystem.conf b/src/collectors/ebpf.plugin/ebpf.d/filesystem.conf
index 209abba77..209abba77 100644
--- a/collectors/ebpf.plugin/ebpf.d/filesystem.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/filesystem.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/functions.conf b/src/collectors/ebpf.plugin/ebpf.d/functions.conf
index a4f57f641..a4f57f641 100644
--- a/collectors/ebpf.plugin/ebpf.d/functions.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/functions.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/hardirq.conf b/src/collectors/ebpf.plugin/ebpf.d/hardirq.conf
index 6a47a94bf..6a47a94bf 100644
--- a/collectors/ebpf.plugin/ebpf.d/hardirq.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/hardirq.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/mdflush.conf b/src/collectors/ebpf.plugin/ebpf.d/mdflush.conf
index ea97ebe85..ea97ebe85 100644
--- a/collectors/ebpf.plugin/ebpf.d/mdflush.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/mdflush.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/mount.conf b/src/collectors/ebpf.plugin/ebpf.d/mount.conf
index ff9a2948c..ff9a2948c 100644
--- a/collectors/ebpf.plugin/ebpf.d/mount.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/mount.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/network.conf b/src/collectors/ebpf.plugin/ebpf.d/network.conf
index 99c32edc1..99c32edc1 100644
--- a/collectors/ebpf.plugin/ebpf.d/network.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/network.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/oomkill.conf b/src/collectors/ebpf.plugin/ebpf.d/oomkill.conf
index ea97ebe85..ea97ebe85 100644
--- a/collectors/ebpf.plugin/ebpf.d/oomkill.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/oomkill.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/process.conf b/src/collectors/ebpf.plugin/ebpf.d/process.conf
index 150c57920..6f6477003 100644
--- a/collectors/ebpf.plugin/ebpf.d/process.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/process.conf
@@ -26,6 +26,6 @@
# cgroups = no
# update every = 10
# pid table size = 32768
- collect pid = real parent
+ collect pid = all
# maps per core = yes
lifetime = 300
diff --git a/collectors/ebpf.plugin/ebpf.d/shm.conf b/src/collectors/ebpf.plugin/ebpf.d/shm.conf
index 95fb54e0f..0314bdc95 100644
--- a/collectors/ebpf.plugin/ebpf.d/shm.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/shm.conf
@@ -31,6 +31,7 @@
# pid table size = 32768
ebpf type format = auto
ebpf co-re tracing = trampoline
+ collect pid = all
# maps per core = yes
lifetime = 300
diff --git a/collectors/ebpf.plugin/ebpf.d/softirq.conf b/src/collectors/ebpf.plugin/ebpf.d/softirq.conf
index 6a47a94bf..6a47a94bf 100644
--- a/collectors/ebpf.plugin/ebpf.d/softirq.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/softirq.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/sync.conf b/src/collectors/ebpf.plugin/ebpf.d/sync.conf
index a086ed4db..a086ed4db 100644
--- a/collectors/ebpf.plugin/ebpf.d/sync.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/sync.conf
diff --git a/collectors/ebpf.plugin/ebpf.d/vfs.conf b/src/collectors/ebpf.plugin/ebpf.d/vfs.conf
index f511581b8..f511581b8 100644
--- a/collectors/ebpf.plugin/ebpf.d/vfs.conf
+++ b/src/collectors/ebpf.plugin/ebpf.d/vfs.conf
diff --git a/collectors/ebpf.plugin/ebpf_apps.c b/src/collectors/ebpf.plugin/ebpf_apps.c
index 10c452267..a17cdb33d 100644
--- a/collectors/ebpf.plugin/ebpf_apps.c
+++ b/src/collectors/ebpf.plugin/ebpf_apps.c
@@ -7,24 +7,6 @@
// ----------------------------------------------------------------------------
// ARAL vectors used to speed up processing
ARAL *ebpf_aral_apps_pid_stat = NULL;
-ARAL *ebpf_aral_process_stat = NULL;
-ARAL *ebpf_aral_socket_pid = NULL;
-ARAL *ebpf_aral_cachestat_pid = NULL;
-ARAL *ebpf_aral_dcstat_pid = NULL;
-ARAL *ebpf_aral_vfs_pid = NULL;
-ARAL *ebpf_aral_fd_pid = NULL;
-ARAL *ebpf_aral_shm_pid = NULL;
-
-// ----------------------------------------------------------------------------
-// Global vectors used with apps
-ebpf_socket_publish_apps_t **socket_bandwidth_curr = NULL;
-netdata_publish_cachestat_t **cachestat_pid = NULL;
-netdata_publish_dcstat_t **dcstat_pid = NULL;
-netdata_publish_swap_t **swap_pid = NULL;
-netdata_publish_vfs_t **vfs_pid = NULL;
-netdata_fd_stat_t **fd_pid = NULL;
-netdata_publish_shm_t **shm_pid = NULL;
-ebpf_process_stat_t **global_process_stats = NULL;
/**
* eBPF ARAL Init
@@ -41,8 +23,6 @@ void ebpf_aral_init(void)
ebpf_aral_apps_pid_stat = ebpf_allocate_pid_aral("ebpf_pid_stat", sizeof(struct ebpf_pid_stat));
- ebpf_aral_process_stat = ebpf_allocate_pid_aral(NETDATA_EBPF_PROC_ARAL_NAME, sizeof(ebpf_process_stat_t));
-
#ifdef NETDATA_DEV_MODE
netdata_log_info("Plugin is using ARAL with values %d", NETDATA_EBPF_ALLOC_MAX_PID);
#endif
@@ -72,266 +52,6 @@ void ebpf_pid_stat_release(struct ebpf_pid_stat *stat)
aral_freez(ebpf_aral_apps_pid_stat, stat);
}
-/*****************************************************************
- *
- * PROCESS ARAL FUNCTIONS
- *
- *****************************************************************/
-
-/**
- * eBPF process stat get
- *
- * Get a ebpf_pid_stat entry to be used with a specific PID.
- *
- * @return it returns the address on success.
- */
-ebpf_process_stat_t *ebpf_process_stat_get(void)
-{
- ebpf_process_stat_t *target = aral_mallocz(ebpf_aral_process_stat);
- memset(target, 0, sizeof(ebpf_process_stat_t));
- return target;
-}
-
-/**
- * eBPF process release
- *
- * @param stat Release a target after usage.
- */
-void ebpf_process_stat_release(ebpf_process_stat_t *stat)
-{
- aral_freez(ebpf_aral_process_stat, stat);
-}
-
-/*****************************************************************
- *
- * SOCKET ARAL FUNCTIONS
- *
- *****************************************************************/
-
-/**
- * eBPF socket Aral init
- *
- * Initiallize array allocator that will be used when integration with apps is enabled.
- */
-void ebpf_socket_aral_init()
-{
- ebpf_aral_socket_pid = ebpf_allocate_pid_aral(NETDATA_EBPF_SOCKET_ARAL_NAME, sizeof(ebpf_socket_publish_apps_t));
-}
-
-/**
- * eBPF socket get
- *
- * Get a ebpf_socket_publish_apps_t entry to be used with a specific PID.
- *
- * @return it returns the address on success.
- */
-ebpf_socket_publish_apps_t *ebpf_socket_stat_get(void)
-{
- ebpf_socket_publish_apps_t *target = aral_mallocz(ebpf_aral_socket_pid);
- memset(target, 0, sizeof(ebpf_socket_publish_apps_t));
- return target;
-}
-
-/*****************************************************************
- *
- * CACHESTAT ARAL FUNCTIONS
- *
- *****************************************************************/
-
-/**
- * eBPF Cachestat Aral init
- *
- * Initiallize array allocator that will be used when integration with apps is enabled.
- */
-void ebpf_cachestat_aral_init()
-{
- ebpf_aral_cachestat_pid = ebpf_allocate_pid_aral(NETDATA_EBPF_CACHESTAT_ARAL_NAME, sizeof(netdata_publish_cachestat_t));
-}
-
-/**
- * eBPF publish cachestat get
- *
- * Get a netdata_publish_cachestat_t entry to be used with a specific PID.
- *
- * @return it returns the address on success.
- */
-netdata_publish_cachestat_t *ebpf_publish_cachestat_get(void)
-{
- netdata_publish_cachestat_t *target = aral_mallocz(ebpf_aral_cachestat_pid);
- memset(target, 0, sizeof(netdata_publish_cachestat_t));
- return target;
-}
-
-/**
- * eBPF cachestat release
- *
- * @param stat Release a target after usage.
- */
-void ebpf_cachestat_release(netdata_publish_cachestat_t *stat)
-{
- aral_freez(ebpf_aral_cachestat_pid, stat);
-}
-
-/*****************************************************************
- *
- * DCSTAT ARAL FUNCTIONS
- *
- *****************************************************************/
-
-/**
- * eBPF directory cache Aral init
- *
- * Initiallize array allocator that will be used when integration with apps is enabled.
- */
-void ebpf_dcstat_aral_init()
-{
- ebpf_aral_dcstat_pid = ebpf_allocate_pid_aral(NETDATA_EBPF_DCSTAT_ARAL_NAME, sizeof(netdata_publish_dcstat_t));
-}
-
-/**
- * eBPF publish dcstat get
- *
- * Get a netdata_publish_dcstat_t entry to be used with a specific PID.
- *
- * @return it returns the address on success.
- */
-netdata_publish_dcstat_t *ebpf_publish_dcstat_get(void)
-{
- netdata_publish_dcstat_t *target = aral_mallocz(ebpf_aral_dcstat_pid);
- memset(target, 0, sizeof(netdata_publish_dcstat_t));
- return target;
-}
-
-/**
- * eBPF dcstat release
- *
- * @param stat Release a target after usage.
- */
-void ebpf_dcstat_release(netdata_publish_dcstat_t *stat)
-{
- aral_freez(ebpf_aral_dcstat_pid, stat);
-}
-
-/*****************************************************************
- *
- * VFS ARAL FUNCTIONS
- *
- *****************************************************************/
-
-/**
- * eBPF VFS Aral init
- *
- * Initiallize array allocator that will be used when integration with apps is enabled.
- */
-void ebpf_vfs_aral_init()
-{
- ebpf_aral_vfs_pid = ebpf_allocate_pid_aral(NETDATA_EBPF_VFS_ARAL_NAME, sizeof(netdata_publish_vfs_t));
-}
-
-/**
- * eBPF publish VFS get
- *
- * Get a netdata_publish_vfs_t entry to be used with a specific PID.
- *
- * @return it returns the address on success.
- */
-netdata_publish_vfs_t *ebpf_vfs_get(void)
-{
- netdata_publish_vfs_t *target = aral_mallocz(ebpf_aral_vfs_pid);
- memset(target, 0, sizeof(netdata_publish_vfs_t));
- return target;
-}
-
-/**
- * eBPF VFS release
- *
- * @param stat Release a target after usage.
- */
-void ebpf_vfs_release(netdata_publish_vfs_t *stat)
-{
- aral_freez(ebpf_aral_vfs_pid, stat);
-}
-
-/*****************************************************************
- *
- * FD ARAL FUNCTIONS
- *
- *****************************************************************/
-
-/**
- * eBPF file descriptor Aral init
- *
- * Initiallize array allocator that will be used when integration with apps is enabled.
- */
-void ebpf_fd_aral_init()
-{
- ebpf_aral_fd_pid = ebpf_allocate_pid_aral(NETDATA_EBPF_FD_ARAL_NAME, sizeof(netdata_fd_stat_t));
-}
-
-/**
- * eBPF publish file descriptor get
- *
- * Get a netdata_fd_stat_t entry to be used with a specific PID.
- *
- * @return it returns the address on success.
- */
-netdata_fd_stat_t *ebpf_fd_stat_get(void)
-{
- netdata_fd_stat_t *target = aral_mallocz(ebpf_aral_fd_pid);
- memset(target, 0, sizeof(netdata_fd_stat_t));
- return target;
-}
-
-/**
- * eBPF file descriptor release
- *
- * @param stat Release a target after usage.
- */
-void ebpf_fd_release(netdata_fd_stat_t *stat)
-{
- aral_freez(ebpf_aral_fd_pid, stat);
-}
-
-/*****************************************************************
- *
- * SHM ARAL FUNCTIONS
- *
- *****************************************************************/
-
-/**
- * eBPF shared memory Aral init
- *
- * Initiallize array allocator that will be used when integration with apps is enabled.
- */
-void ebpf_shm_aral_init()
-{
- ebpf_aral_shm_pid = ebpf_allocate_pid_aral(NETDATA_EBPF_SHM_ARAL_NAME, sizeof(netdata_publish_shm_t));
-}
-
-/**
- * eBPF shared memory get
- *
- * Get a netdata_publish_shm_t entry to be used with a specific PID.
- *
- * @return it returns the address on success.
- */
-netdata_publish_shm_t *ebpf_shm_stat_get(void)
-{
- netdata_publish_shm_t *target = aral_mallocz(ebpf_aral_shm_pid);
- memset(target, 0, sizeof(netdata_publish_shm_t));
- return target;
-}
-
-/**
- * eBPF shared memory release
- *
- * @param stat Release a target after usage.
- */
-void ebpf_shm_release(netdata_publish_shm_t *stat)
-{
- aral_freez(ebpf_aral_shm_pid, stat);
-}
-
// ----------------------------------------------------------------------------
// internal flags
// handled in code (automatically set)
@@ -372,24 +92,6 @@ int ebpf_read_hash_table(void *ep, int fd, uint32_t pid)
*****************************************************************/
/**
- * Am I running as Root
- *
- * Verify the user that is running the collector.
- *
- * @return It returns 1 for root and 0 otherwise.
- */
-int am_i_running_as_root()
-{
- uid_t uid = getuid(), euid = geteuid();
-
- if (uid == 0 || euid == 0) {
- return 1;
- }
-
- return 0;
-}
-
-/**
* Reset the target values
*
* @param root the pointer to the chain that will be reset.
@@ -753,14 +455,19 @@ static inline int managed_log(struct ebpf_pid_stat *p, uint32_t log, int status)
*
* Get or allocate the PID entry for the specified pid.
*
- * @param pid the pid to search the data.
+ * @param pid the pid to search the data.
+ * @param tgid the task group id
*
* @return It returns the pid entry structure
*/
-static inline struct ebpf_pid_stat *get_pid_entry(pid_t pid)
+ebpf_pid_stat_t *ebpf_get_pid_entry(pid_t pid, pid_t tgid)
{
- if (unlikely(ebpf_all_pids[pid]))
+ ebpf_pid_stat_t *ptr = ebpf_all_pids[pid];
+ if (unlikely(ptr)) {
+ if (!ptr->ppid && tgid)
+ ptr->ppid = tgid;
return ebpf_all_pids[pid];
+ }
struct ebpf_pid_stat *p = ebpf_pid_stat_get();
@@ -771,6 +478,7 @@ static inline struct ebpf_pid_stat *get_pid_entry(pid_t pid)
ebpf_root_of_pids = p;
p->pid = pid;
+ p->ppid = tgid;
ebpf_all_pids[pid] = p;
ebpf_all_pids_count++;
@@ -951,14 +659,14 @@ static inline int read_proc_pid_stat(struct ebpf_pid_stat *p, void *ptr)
*
* @return It returns 1 on success and 0 otherwise
*/
-static inline int collect_data_for_pid(pid_t pid, void *ptr)
+static inline int ebpf_collect_data_for_pid(pid_t pid, void *ptr)
{
if (unlikely(pid < 0 || pid > pid_max)) {
netdata_log_error("Invalid pid %d read (expected %d to %d). Ignoring process.", pid, 0, pid_max);
return 0;
}
- struct ebpf_pid_stat *p = get_pid_entry(pid);
+ ebpf_pid_stat_t *p = ebpf_get_pid_entry(pid, 0);
if (unlikely(!p || p->read))
return 0;
p->read = 1;
@@ -1164,7 +872,7 @@ static inline void post_aggregate_targets(struct ebpf_target *root)
*
* @param pid the PID that will be removed.
*/
-static inline void del_pid_entry(pid_t pid)
+static inline void ebpf_del_pid_entry(pid_t pid)
{
struct ebpf_pid_stat *p = ebpf_all_pids[pid];
@@ -1201,6 +909,7 @@ static inline void del_pid_entry(pid_t pid)
}
JudyLFreeArray(&pid_ptr->socket_stats.JudyLArray, PJE0);
}
+ aral_freez(ebpf_judy_pid.pid_table, pid_ptr);
JudyLDel(&ebpf_judy_pid.index.JudyLArray, p->pid, PJE0);
}
rw_spinlock_write_unlock(&ebpf_judy_pid.index.rw_spinlock);
@@ -1240,79 +949,23 @@ int get_pid_comm(pid_t pid, size_t n, char *dest)
}
/**
- * Cleanup variable from other threads
- *
- * @param pid current pid.
- */
-void cleanup_variables_from_other_threads(uint32_t pid)
-{
- // Clean cachestat structure
- if (cachestat_pid) {
- ebpf_cachestat_release(cachestat_pid[pid]);
- cachestat_pid[pid] = NULL;
- }
-
- // Clean directory cache structure
- if (dcstat_pid) {
- ebpf_dcstat_release(dcstat_pid[pid]);
- dcstat_pid[pid] = NULL;
- }
-
- // Clean swap structure
- if (swap_pid) {
- freez(swap_pid[pid]);
- swap_pid[pid] = NULL;
- }
-
- // Clean vfs structure
- if (vfs_pid) {
- ebpf_vfs_release(vfs_pid[pid]);
- vfs_pid[pid] = NULL;
- }
-
- // Clean fd structure
- if (fd_pid) {
- ebpf_fd_release(fd_pid[pid]);
- fd_pid[pid] = NULL;
- }
-
- // Clean shm structure
- if (shm_pid) {
- ebpf_shm_release(shm_pid[pid]);
- shm_pid[pid] = NULL;
- }
-}
-
-/**
* Remove PIDs when they are not running more.
*/
-void cleanup_exited_pids()
+void ebpf_cleanup_exited_pids(int max)
{
struct ebpf_pid_stat *p = NULL;
for (p = ebpf_root_of_pids; p;) {
- if (!p->updated && (!p->keep || p->keeploops > 0)) {
+ if (p->not_updated > max) {
if (unlikely(debug_enabled && (p->keep || p->keeploops)))
debug_log(" > CLEANUP cannot keep exited process %d (%s) anymore - removing it.", p->pid, p->comm);
pid_t r = p->pid;
p = p->next;
- // Clean process structure
- if (global_process_stats) {
- ebpf_process_stat_release(global_process_stats[r]);
- global_process_stats[r] = NULL;
- }
-
- cleanup_variables_from_other_threads(r);
-
- del_pid_entry(r);
- } else {
- if (unlikely(p->keep))
- p->keeploops++;
- p->keep = 0;
- p = p->next;
+ ebpf_del_pid_entry(r);
}
+ p = p->next;
}
}
@@ -1344,7 +997,7 @@ static inline void read_proc_filesystem()
if (unlikely(endptr == de->d_name || *endptr != '\0'))
continue;
- collect_data_for_pid(pid, NULL);
+ ebpf_collect_data_for_pid(pid, NULL);
}
closedir(dir);
}
@@ -1400,6 +1053,31 @@ void ebpf_process_apps_accumulator(ebpf_process_stat_t *out, int maps_per_core)
}
/**
+ * Sum values for pid
+ *
+ * @param structure to store result.
+ * @param root the structure with all available PIDs
+ */
+void ebpf_process_sum_values_for_pids(ebpf_process_stat_t *process, struct ebpf_pid_on_target *root)
+{
+ memset(process, 0, sizeof(ebpf_process_stat_t));
+ while (root) {
+ int32_t pid = root->pid;
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ ebpf_process_stat_t *in = &local_pid->process;
+ process->task_err += in->task_err;
+ process->release_call += in->release_call;
+ process->exit_call += in->exit_call;
+ process->create_thread += in->create_thread;
+ process->create_process += in->create_process;
+ }
+
+ root = root->next;
+ }
+}
+
+/**
* Collect data for all process
*
* Read data from hash table and store it in appropriate vectors.
@@ -1431,42 +1109,31 @@ void collect_data_for_all_processes(int tbl_pid_stats_fd, int maps_per_core)
read_proc_filesystem();
- uint32_t key;
pids = ebpf_root_of_pids; // global list of all processes running
- // while (bpf_map_get_next_key(tbl_pid_stats_fd, &key, &next_key) == 0) {
if (tbl_pid_stats_fd != -1) {
size_t length = sizeof(ebpf_process_stat_t);
if (maps_per_core)
length *= ebpf_nprocs;
- while (pids) {
- key = pids->pid;
-
- ebpf_process_stat_t *w = global_process_stats[key];
- if (!w) {
- w = ebpf_process_stat_get();
- global_process_stats[key] = w;
- }
+ uint32_t key = 0, next_key = 0;
+ while (bpf_map_get_next_key(tbl_pid_stats_fd, &key, &next_key) == 0) {
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(key, 0);
+ if (!local_pid)
+ goto end_process_loop;
+ ebpf_process_stat_t *w = &local_pid->process;
if (bpf_map_lookup_elem(tbl_pid_stats_fd, &key, process_stat_vector)) {
- // Clean Process structures
- ebpf_process_stat_release(w);
- global_process_stats[key] = NULL;
-
- cleanup_variables_from_other_threads(key);
-
- pids = pids->next;
- continue;
+ goto end_process_loop;
}
ebpf_process_apps_accumulator(process_stat_vector, maps_per_core);
memcpy(w, process_stat_vector, sizeof(ebpf_process_stat_t));
+end_process_loop:
memset(process_stat_vector, 0, length);
-
- pids = pids->next;
+ key = next_key;
}
}
@@ -1482,4 +1149,12 @@ void collect_data_for_all_processes(int tbl_pid_stats_fd, int maps_per_core)
aggregate_pid_on_target(pids->target, pids, NULL);
post_aggregate_targets(apps_groups_root_target);
+
+ struct ebpf_target *w;
+ for (w = apps_groups_root_target; w; w = w->next) {
+ if (unlikely(!(w->processes)))
+ continue;
+
+ ebpf_process_sum_values_for_pids(&w->process, w->root_pid);
+ }
}
diff --git a/collectors/ebpf.plugin/ebpf_apps.h b/src/collectors/ebpf.plugin/ebpf_apps.h
index 258091507..a2cbaf3b7 100644
--- a/collectors/ebpf.plugin/ebpf_apps.h
+++ b/src/collectors/ebpf.plugin/ebpf_apps.h
@@ -13,11 +13,14 @@
#define NETDATA_APP_FAMILY "app"
#define NETDATA_APPS_FILE_GROUP "file_access"
#define NETDATA_APPS_FILE_FDS "fds"
-#define NETDATA_APPS_FILE_CGROUP_GROUP "file_access (eBPF)"
-#define NETDATA_APPS_PROCESS_GROUP "process (eBPF)"
+#define NETDATA_APPS_PROCESS_GROUP "process"
#define NETDATA_APPS_NET_GROUP "net"
#define NETDATA_APPS_IPC_SHM_GROUP "ipc shm"
+#ifndef TASK_COMM_LEN
+#define TASK_COMM_LEN 16
+#endif
+
#include "ebpf_process.h"
#include "ebpf_dcstat.h"
#include "ebpf_disk.h"
@@ -39,6 +42,31 @@
#define EBPF_MAX_COMPARE_NAME 100
#define EBPF_MAX_NAME 100
+#define EBPF_CLEANUP_FACTOR 10
+
+// ----------------------------------------------------------------------------
+// Structures used to read information from kernel ring
+typedef struct ebpf_process_stat {
+ uint64_t ct;
+ uint32_t uid;
+ uint32_t gid;
+ char name[TASK_COMM_LEN];
+
+ uint32_t tgid;
+ uint32_t pid;
+
+ //Counter
+ uint32_t exit_call;
+ uint32_t release_call;
+ uint32_t create_process;
+ uint32_t create_thread;
+
+ //Counter
+ uint32_t task_err;
+
+ uint8_t removeme;
+} ebpf_process_stat_t;
+
// ----------------------------------------------------------------------------
// pid_stat
//
@@ -61,6 +89,8 @@ struct ebpf_target {
netdata_publish_vfs_t vfs;
netdata_fd_stat_t fd;
netdata_publish_shm_t shm;
+ ebpf_process_stat_t process;
+ ebpf_socket_publish_apps_t socket;
kernel_uint_t starttime;
kernel_uint_t collected_starttime;
@@ -84,7 +114,7 @@ extern struct ebpf_target *apps_groups_root_target;
extern struct ebpf_target *users_root_target;
extern struct ebpf_target *groups_root_target;
-struct ebpf_pid_stat {
+typedef struct ebpf_pid_stat {
int32_t pid;
char comm[EBPF_MAX_COMPARE_NAME + 1];
char *cmdline;
@@ -105,6 +135,16 @@ struct ebpf_pid_stat {
int sortlist; // higher numbers = top on the process tree
// each process gets a unique number
+ netdata_publish_cachestat_t cachestat;
+ netdata_publish_dcstat_t dc;
+ netdata_fd_stat_t fd;
+ ebpf_process_stat_t process;
+ netdata_publish_shm_t shm;
+ netdata_publish_swap_t swap;
+ ebpf_socket_publish_apps_t socket;
+ netdata_publish_vfs_t vfs;
+
+ int not_updated;
struct ebpf_target *target; // app_groups.conf targets
struct ebpf_target *user_target; // uid based targets
@@ -113,6 +153,8 @@ struct ebpf_pid_stat {
usec_t stat_collected_usec;
usec_t last_stat_collected_usec;
+ netdata_publish_cachestat_t cache;
+
char *stat_filename;
char *status_filename;
char *io_filename;
@@ -121,7 +163,7 @@ struct ebpf_pid_stat {
struct ebpf_pid_stat *parent;
struct ebpf_pid_stat *prev;
struct ebpf_pid_stat *next;
-};
+} ebpf_pid_stat_t;
// ----------------------------------------------------------------------------
// target
@@ -136,24 +178,6 @@ struct ebpf_pid_on_target {
struct ebpf_pid_on_target *next;
};
-// ----------------------------------------------------------------------------
-// Structures used to read information from kernel ring
-typedef struct ebpf_process_stat {
- uint64_t pid_tgid; // This cannot be removed, because it is used inside kernel ring.
- uint32_t pid;
-
- //Counter
- uint32_t exit_call;
- uint32_t release_call;
- uint32_t create_process;
- uint32_t create_thread;
-
- //Counter
- uint32_t task_err;
-
- uint8_t removeme;
-} ebpf_process_stat_t;
-
/**
* Internal function used to write debug messages.
*
@@ -186,8 +210,6 @@ void clean_apps_groups_target(struct ebpf_target *apps_groups_root_target);
size_t zero_all_targets(struct ebpf_target *root);
-int am_i_running_as_root();
-
void cleanup_exited_pids();
int ebpf_read_hash_table(void *ep, int fd, uint32_t pid);
@@ -197,16 +219,8 @@ int get_pid_comm(pid_t pid, size_t n, char *dest);
void collect_data_for_all_processes(int tbl_pid_stats_fd, int maps_per_core);
void ebpf_process_apps_accumulator(ebpf_process_stat_t *out, int maps_per_core);
-extern ebpf_process_stat_t **global_process_stats;
-extern netdata_publish_cachestat_t **cachestat_pid;
-extern netdata_publish_dcstat_t **dcstat_pid;
-extern netdata_publish_swap_t **swap_pid;
-extern netdata_publish_vfs_t **vfs_pid;
-extern netdata_fd_stat_t **fd_pid;
-extern netdata_publish_shm_t **shm_pid;
-
// The default value is at least 32 times smaller than maximum number of PIDs allowed on system,
-// this is only possible because we are using ARAL (https://github.com/netdata/netdata/tree/master/libnetdata/aral).
+// this is only possible because we are using ARAL (https://github.com/netdata/netdata/tree/master/src/libnetdata/aral).
#ifndef NETDATA_EBPF_ALLOC_MAX_PID
# define NETDATA_EBPF_ALLOC_MAX_PID 1024
#endif
@@ -214,51 +228,25 @@ extern netdata_publish_shm_t **shm_pid;
// ARAL Sectiion
extern void ebpf_aral_init(void);
-
-extern ebpf_process_stat_t *ebpf_process_stat_get(void);
-extern void ebpf_process_stat_release(ebpf_process_stat_t *stat);
+extern ebpf_pid_stat_t *ebpf_get_pid_entry(pid_t pid, pid_t tgid);
extern ebpf_process_stat_t *process_stat_vector;
-extern ARAL *ebpf_aral_socket_pid;
-void ebpf_socket_aral_init();
-ebpf_socket_publish_apps_t *ebpf_socket_stat_get(void);
-
-extern ARAL *ebpf_aral_cachestat_pid;
-void ebpf_cachestat_aral_init();
-netdata_publish_cachestat_t *ebpf_publish_cachestat_get(void);
-void ebpf_cachestat_release(netdata_publish_cachestat_t *stat);
-
-extern ARAL *ebpf_aral_dcstat_pid;
-void ebpf_dcstat_aral_init();
-netdata_publish_dcstat_t *ebpf_publish_dcstat_get(void);
-void ebpf_dcstat_release(netdata_publish_dcstat_t *stat);
-
extern ARAL *ebpf_aral_vfs_pid;
void ebpf_vfs_aral_init();
netdata_publish_vfs_t *ebpf_vfs_get(void);
void ebpf_vfs_release(netdata_publish_vfs_t *stat);
-extern ARAL *ebpf_aral_fd_pid;
-void ebpf_fd_aral_init();
-netdata_fd_stat_t *ebpf_fd_stat_get(void);
-void ebpf_fd_release(netdata_fd_stat_t *stat);
-
extern ARAL *ebpf_aral_shm_pid;
void ebpf_shm_aral_init();
netdata_publish_shm_t *ebpf_shm_stat_get(void);
void ebpf_shm_release(netdata_publish_shm_t *stat);
+void ebpf_cleanup_exited_pids(int max);
// ARAL Section end
// Threads integrated with apps
-extern ebpf_socket_publish_apps_t **socket_bandwidth_curr;
// Threads integrated with apps
#include "libnetdata/threads/threads.h"
-// ARAL variables
-extern ARAL *ebpf_aral_apps_pid_stat;
-extern ARAL *ebpf_aral_process_stat;
-#define NETDATA_EBPF_PROC_ARAL_NAME "ebpf_proc_stat"
-
#endif /* NETDATA_EBPF_APPS_H */
diff --git a/collectors/ebpf.plugin/ebpf_cachestat.c b/src/collectors/ebpf.plugin/ebpf_cachestat.c
index d9f8f7b06..45e09766f 100644
--- a/collectors/ebpf.plugin/ebpf_cachestat.c
+++ b/src/collectors/ebpf.plugin/ebpf_cachestat.c
@@ -58,9 +58,16 @@ netdata_ebpf_targets_t cachestat_targets[] = { {.name = "add_to_page_cache_lru",
static char *account_page[NETDATA_CACHESTAT_ACCOUNT_DIRTY_END] ={ "account_page_dirtied",
"__set_page_dirty", "__folio_mark_dirty" };
-#ifdef NETDATA_DEV_MODE
-int cachestat_disable_priority;
-#endif
+struct netdata_static_thread ebpf_read_cachestat = {
+ .name = "EBPF_READ_CACHESTAT",
+ .config_section = NULL,
+ .config_name = NULL,
+ .env_name = NULL,
+ .enabled = 1,
+ .thread = NULL,
+ .init_routine = NULL,
+ .start_routine = NULL
+};
#ifdef LIBBPF_MAJOR_VERSION
/**
@@ -78,7 +85,6 @@ static void ebpf_cachestat_disable_probe(struct cachestat_bpf *obj)
bpf_program__set_autoload(obj->progs.netdata_set_page_dirty_kprobe, false);
bpf_program__set_autoload(obj->progs.netdata_account_page_dirtied_kprobe, false);
bpf_program__set_autoload(obj->progs.netdata_mark_buffer_dirty_kprobe, false);
- bpf_program__set_autoload(obj->progs.netdata_release_task_kprobe, false);
}
/*
@@ -119,7 +125,6 @@ static void ebpf_cachestat_disable_trampoline(struct cachestat_bpf *obj)
bpf_program__set_autoload(obj->progs.netdata_set_page_dirty_fentry, false);
bpf_program__set_autoload(obj->progs.netdata_account_page_dirtied_fentry, false);
bpf_program__set_autoload(obj->progs.netdata_mark_buffer_dirty_fentry, false);
- bpf_program__set_autoload(obj->progs.netdata_release_task_fentry, false);
}
/*
@@ -175,9 +180,6 @@ static inline void netdata_set_trampoline_target(struct cachestat_bpf *obj)
bpf_program__set_attach_target(obj->progs.netdata_mark_buffer_dirty_fentry, 0,
cachestat_targets[NETDATA_KEY_CALLS_MARK_BUFFER_DIRTY].name);
-
- bpf_program__set_attach_target(obj->progs.netdata_release_task_fentry, 0,
- EBPF_COMMON_FNCT_CLEAN_UP);
}
/**
@@ -194,7 +196,7 @@ static int ebpf_cachestat_attach_probe(struct cachestat_bpf *obj)
obj->links.netdata_add_to_page_cache_lru_kprobe = bpf_program__attach_kprobe(obj->progs.netdata_add_to_page_cache_lru_kprobe,
false,
cachestat_targets[NETDATA_KEY_CALLS_ADD_TO_PAGE_CACHE_LRU].name);
- int ret = libbpf_get_error(obj->links.netdata_add_to_page_cache_lru_kprobe);
+ long ret = libbpf_get_error(obj->links.netdata_add_to_page_cache_lru_kprobe);
if (ret)
return -1;
@@ -234,13 +236,6 @@ static int ebpf_cachestat_attach_probe(struct cachestat_bpf *obj)
if (ret)
return -1;
- obj->links.netdata_release_task_kprobe = bpf_program__attach_kprobe(obj->progs.netdata_release_task_kprobe,
- false,
- EBPF_COMMON_FNCT_CLEAN_UP);
- ret = libbpf_get_error(obj->links.netdata_release_task_kprobe);
- if (ret)
- return -1;
-
return 0;
}
@@ -277,19 +272,6 @@ static void ebpf_cachestat_set_hash_tables(struct cachestat_bpf *obj)
}
/**
- * Disable Release Task
- *
- * Disable release task when apps is not enabled.
- *
- * @param obj is the main structure for bpf objects.
- */
-static void ebpf_cachestat_disable_release_task(struct cachestat_bpf *obj)
-{
- bpf_program__set_autoload(obj->progs.netdata_release_task_kprobe, false);
- bpf_program__set_autoload(obj->progs.netdata_release_task_fentry, false);
-}
-
-/**
* Load and attach
*
* Load and attach the eBPF code in kernel.
@@ -316,9 +298,6 @@ static inline int ebpf_cachestat_load_and_attach(struct cachestat_bpf *obj, ebpf
ebpf_cachestat_adjust_map(obj, em);
- if (!em->apps_charts && !em->cgroup_charts)
- ebpf_cachestat_disable_release_task(obj);
-
int ret = cachestat_bpf__load(obj);
if (ret) {
return ret;
@@ -349,11 +328,11 @@ static void ebpf_obsolete_specific_cachestat_charts(char *type, int update_every
*
* @param em a pointer to `struct ebpf_module`
*/
-static void ebpf_obsolete_services(ebpf_module_t *em)
+static void ebpf_obsolete_cachestat_services(ebpf_module_t *em, char *id)
{
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_CACHESTAT_HIT_RATIO_CHART,
- "",
"Hit ratio",
EBPF_COMMON_DIMENSION_PERCENTAGE,
NETDATA_CACHESTAT_SUBMENU,
@@ -363,8 +342,8 @@ static void ebpf_obsolete_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_CACHESTAT_DIRTY_CHART,
- "",
"Number of dirty pages",
EBPF_CACHESTAT_DIMENSION_PAGE,
NETDATA_CACHESTAT_SUBMENU,
@@ -374,8 +353,8 @@ static void ebpf_obsolete_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_CACHESTAT_HIT_CHART,
- "",
"Number of accessed files",
EBPF_CACHESTAT_DIMENSION_HITS,
NETDATA_CACHESTAT_SUBMENU,
@@ -385,8 +364,8 @@ static void ebpf_obsolete_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_CACHESTAT_MISSES_CHART,
- "",
"Files out of page cache",
EBPF_CACHESTAT_DIMENSION_MISSES,
NETDATA_CACHESTAT_SUBMENU,
@@ -406,12 +385,13 @@ static void ebpf_obsolete_services(ebpf_module_t *em)
static inline void ebpf_obsolete_cachestat_cgroup_charts(ebpf_module_t *em) {
pthread_mutex_lock(&mutex_cgroup_shm);
- ebpf_obsolete_services(em);
-
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (ect->systemd)
+ if (ect->systemd) {
+ ebpf_obsolete_cachestat_services(em, ect->name);
+
continue;
+ }
ebpf_obsolete_specific_cachestat_charts(ect->name, em->update_every);
}
@@ -483,6 +463,7 @@ void ebpf_obsolete_cachestat_apps_charts(struct ebpf_module *em)
{
struct ebpf_target *w;
int update_every = em->update_every;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = apps_groups_root_target; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_CACHESTAT_IDX))))
continue;
@@ -532,6 +513,7 @@ void ebpf_obsolete_cachestat_apps_charts(struct ebpf_module *em)
update_every);
w->charts_created &= ~(1<<EBPF_MODULE_CACHESTAT_IDX);
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -545,6 +527,9 @@ static void ebpf_cachestat_exit(void *ptr)
{
ebpf_module_t *em = (ebpf_module_t *)ptr;
+ if (ebpf_read_cachestat.thread)
+ netdata_thread_cancel(*ebpf_read_cachestat.thread);
+
if (em->enabled == NETDATA_THREAD_EBPF_FUNCTION_RUNNING) {
pthread_mutex_lock(&lock);
if (em->cgroup_charts) {
@@ -558,12 +543,6 @@ static void ebpf_cachestat_exit(void *ptr)
ebpf_obsolete_cachestat_global(em);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_cachestat_pid)
- ebpf_statistic_obsolete_aral_chart(em, cachestat_disable_priority);
-#endif
-
-
fflush(stdout);
pthread_mutex_unlock(&lock);
}
@@ -688,13 +667,17 @@ static void cachestat_apps_accumulator(netdata_cachestat_pid_t *out, int maps_pe
{
int i, end = (maps_per_core) ? ebpf_nprocs : 1;
netdata_cachestat_pid_t *total = &out[0];
+ uint64_t ct = total->ct;
for (i = 1; i < end; i++) {
netdata_cachestat_pid_t *w = &out[i];
total->account_page_dirtied += w->account_page_dirtied;
total->add_to_page_cache_lru += w->add_to_page_cache_lru;
total->mark_buffer_dirty += w->mark_buffer_dirty;
total->mark_page_accessed += w->mark_page_accessed;
+ if (w->ct > ct)
+ ct = w->ct;
}
+ total->ct = ct;
}
/**
@@ -703,39 +686,18 @@ static void cachestat_apps_accumulator(netdata_cachestat_pid_t *out, int maps_pe
* Save the current values inside the structure
*
* @param out vector used to plot charts
- * @param publish vector with values read from hash tables.
+ * @param in vector with values read from hash tables.
*/
-static inline void cachestat_save_pid_values(netdata_publish_cachestat_t *out, netdata_cachestat_pid_t *publish)
+static inline void cachestat_save_pid_values(netdata_publish_cachestat_t *out, netdata_cachestat_pid_t *in)
{
+ out->ct = in->ct;
if (!out->current.mark_page_accessed) {
- memcpy(&out->current, &publish[0], sizeof(netdata_cachestat_pid_t));
+ memcpy(&out->current, &in[0], sizeof(netdata_cachestat_pid_t));
return;
}
memcpy(&out->prev, &out->current, sizeof(netdata_cachestat_pid_t));
- memcpy(&out->current, &publish[0], sizeof(netdata_cachestat_pid_t));
-}
-
-/**
- * Fill PID
- *
- * Fill PID structures
- *
- * @param current_pid pid that we are collecting data
- * @param out values read from hash tables;
- */
-static void cachestat_fill_pid(uint32_t current_pid, netdata_cachestat_pid_t *publish)
-{
- netdata_publish_cachestat_t *curr = cachestat_pid[current_pid];
- if (!curr) {
- curr = ebpf_publish_cachestat_get();
- cachestat_pid[current_pid] = curr;
-
- cachestat_save_pid_values(curr, publish);
- return;
- }
-
- cachestat_save_pid_values(curr, publish);
+ memcpy(&out->current, &in[0], sizeof(netdata_cachestat_pid_t));
}
/**
@@ -745,32 +707,39 @@ static void cachestat_fill_pid(uint32_t current_pid, netdata_cachestat_pid_t *pu
*
* @param maps_per_core do I need to read all cores?
*/
-static void ebpf_read_cachestat_apps_table(int maps_per_core)
+static void ebpf_read_cachestat_apps_table(int maps_per_core, int max_period)
{
netdata_cachestat_pid_t *cv = cachestat_vector;
- uint32_t key;
- struct ebpf_pid_stat *pids = ebpf_root_of_pids;
int fd = cachestat_maps[NETDATA_CACHESTAT_PID_STATS].map_fd;
size_t length = sizeof(netdata_cachestat_pid_t);
if (maps_per_core)
length *= ebpf_nprocs;
- while (pids) {
- key = pids->pid;
-
+ uint32_t key = 0, next_key = 0;
+ while (bpf_map_get_next_key(fd, &key, &next_key) == 0) {
if (bpf_map_lookup_elem(fd, &key, cv)) {
- pids = pids->next;
- continue;
+ goto end_cachestat_loop;
}
cachestat_apps_accumulator(cv, maps_per_core);
- cachestat_fill_pid(key, cv);
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(key, cv->tgid);
+ if (!local_pid)
+ goto end_cachestat_loop;
+
+ netdata_publish_cachestat_t *publish = &local_pid->cachestat;
+ if (!publish->ct || publish->ct != cv->ct){
+ cachestat_save_pid_values(publish, cv);
+ local_pid->not_updated = 0;
+ } else if (++local_pid->not_updated >= max_period) {
+ bpf_map_delete_elem(fd, &key);
+ local_pid->not_updated = 0;
+ }
+end_cachestat_loop:
// We are cleaning to avoid passing data read from one process to other.
memset(cv, 0, length);
-
- pids = pids->next;
+ key = next_key;
}
}
@@ -781,14 +750,8 @@ static void ebpf_read_cachestat_apps_table(int maps_per_core)
*
* @param maps_per_core do I need to read all cores?
*/
-static void ebpf_update_cachestat_cgroup(int maps_per_core)
+static void ebpf_update_cachestat_cgroup()
{
- netdata_cachestat_pid_t *cv = cachestat_vector;
- int fd = cachestat_maps[NETDATA_CACHESTAT_PID_STATS].map_fd;
- size_t length = sizeof(netdata_cachestat_pid_t);
- if (maps_per_core)
- length *= ebpf_nprocs;
-
ebpf_cgroup_target_t *ect;
pthread_mutex_lock(&mutex_cgroup_shm);
for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
@@ -796,19 +759,11 @@ static void ebpf_update_cachestat_cgroup(int maps_per_core)
for (pids = ect->pids; pids; pids = pids->next) {
int pid = pids->pid;
netdata_cachestat_pid_t *out = &pids->cachestat;
- if (likely(cachestat_pid) && cachestat_pid[pid]) {
- netdata_publish_cachestat_t *in = cachestat_pid[pid];
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ netdata_publish_cachestat_t *in = &local_pid->cachestat;
memcpy(out, &in->current, sizeof(netdata_cachestat_pid_t));
- } else {
- memset(cv, 0, length);
- if (bpf_map_lookup_elem(fd, &pid, cv)) {
- continue;
- }
-
- cachestat_apps_accumulator(cv, maps_per_core);
-
- memcpy(out, cv, sizeof(netdata_cachestat_pid_t));
}
}
}
@@ -816,6 +771,104 @@ static void ebpf_update_cachestat_cgroup(int maps_per_core)
}
/**
+ * Cachestat sum PIDs
+ *
+ * Sum values for all PIDs associated to a group
+ *
+ * @param publish output structure.
+ * @param root structure with listed IPs
+ */
+void ebpf_cachestat_sum_pids(netdata_publish_cachestat_t *publish, struct ebpf_pid_on_target *root)
+{
+ memcpy(&publish->prev, &publish->current,sizeof(publish->current));
+ memset(&publish->current, 0, sizeof(publish->current));
+
+ netdata_cachestat_pid_t *dst = &publish->current;
+ while (root) {
+ int32_t pid = root->pid;
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ netdata_publish_cachestat_t *w = &local_pid->cachestat;
+ netdata_cachestat_pid_t *src = &w->current;
+ dst->account_page_dirtied += src->account_page_dirtied;
+ dst->add_to_page_cache_lru += src->add_to_page_cache_lru;
+ dst->mark_buffer_dirty += src->mark_buffer_dirty;
+ dst->mark_page_accessed += src->mark_page_accessed;
+ }
+
+ root = root->next;
+ }
+}
+
+/**
+ * Resume apps data
+ */
+void ebpf_resume_apps_data()
+{
+ struct ebpf_target *w;
+
+ for (w = apps_groups_root_target; w; w = w->next) {
+ if (unlikely(!(w->charts_created & (1 << EBPF_MODULE_CACHESTAT_IDX))))
+ continue;
+
+ ebpf_cachestat_sum_pids(&w->cachestat, w->root_pid);
+ }
+}
+
+/**
+ * Cachestat thread
+ *
+ * Thread used to generate cachestat charts.
+ *
+ * @param ptr a pointer to `struct ebpf_module`
+ *
+ * @return It always return NULL
+ */
+void *ebpf_read_cachestat_thread(void *ptr)
+{
+ heartbeat_t hb;
+ heartbeat_init(&hb);
+
+ ebpf_module_t *em = (ebpf_module_t *)ptr;
+
+ int maps_per_core = em->maps_per_core;
+ int update_every = em->update_every;
+ int max_period = update_every * EBPF_CLEANUP_FACTOR;
+
+ int counter = update_every - 1;
+
+ uint32_t lifetime = em->lifetime;
+ uint32_t running_time = 0;
+ usec_t period = update_every * USEC_PER_SEC;
+ while (!ebpf_plugin_exit && running_time < lifetime) {
+ (void)heartbeat_next(&hb, period);
+ if (ebpf_plugin_exit || ++counter != update_every)
+ continue;
+
+ netdata_thread_disable_cancelability();
+
+ pthread_mutex_lock(&collect_data_mutex);
+ ebpf_read_cachestat_apps_table(maps_per_core, max_period);
+ ebpf_resume_apps_data();
+ pthread_mutex_unlock(&collect_data_mutex);
+
+ counter = 0;
+
+ pthread_mutex_lock(&ebpf_exit_cleanup);
+ if (running_time && !em->running_time)
+ running_time = update_every;
+ else
+ running_time += update_every;
+
+ em->running_time = running_time;
+ pthread_mutex_unlock(&ebpf_exit_cleanup);
+ netdata_thread_enable_cancelability();
+ }
+
+ return NULL;
+}
+
+/**
* Create apps charts
*
* Call ebpf_create_chart to create the charts on apps submenu.
@@ -944,7 +997,7 @@ static void cachestat_send_global(netdata_publish_cachestat_t *publish)
ebpf_one_dimension_write_charts(
NETDATA_EBPF_MEMORY_GROUP, NETDATA_CACHESTAT_DIRTY_CHART, ptr[NETDATA_CACHESTAT_IDX_DIRTY].dimension,
- cachestat_hash_values[NETDATA_KEY_CALLS_MARK_BUFFER_DIRTY]);
+ (long long)cachestat_hash_values[NETDATA_KEY_CALLS_MARK_BUFFER_DIRTY]);
ebpf_one_dimension_write_charts(
NETDATA_EBPF_MEMORY_GROUP, NETDATA_CACHESTAT_HIT_CHART, ptr[NETDATA_CACHESTAT_IDX_HIT].dimension, publish->hit);
@@ -955,35 +1008,6 @@ static void cachestat_send_global(netdata_publish_cachestat_t *publish)
}
/**
- * Cachestat sum PIDs
- *
- * Sum values for all PIDs associated to a group
- *
- * @param publish output structure.
- * @param root structure with listed IPs
- */
-void ebpf_cachestat_sum_pids(netdata_publish_cachestat_t *publish, struct ebpf_pid_on_target *root)
-{
- memcpy(&publish->prev, &publish->current,sizeof(publish->current));
- memset(&publish->current, 0, sizeof(publish->current));
-
- netdata_cachestat_pid_t *dst = &publish->current;
- while (root) {
- int32_t pid = root->pid;
- netdata_publish_cachestat_t *w = cachestat_pid[pid];
- if (w) {
- netdata_cachestat_pid_t *src = &w->current;
- dst->account_page_dirtied += src->account_page_dirtied;
- dst->add_to_page_cache_lru += src->add_to_page_cache_lru;
- dst->mark_buffer_dirty += src->mark_buffer_dirty;
- dst->mark_page_accessed += src->mark_page_accessed;
- }
-
- root = root->next;
- }
-}
-
-/**
* Send data to Netdata calling auxiliary functions.
*
* @param root the target list.
@@ -993,17 +1017,17 @@ void ebpf_cache_send_apps_data(struct ebpf_target *root)
struct ebpf_target *w;
collected_number value;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = root; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_CACHESTAT_IDX))))
continue;
- ebpf_cachestat_sum_pids(&w->cachestat, w->root_pid);
netdata_cachestat_pid_t *current = &w->cachestat.current;
netdata_cachestat_pid_t *prev = &w->cachestat.prev;
uint64_t mpa = current->mark_page_accessed - prev->mark_page_accessed;
uint64_t mbd = current->mark_buffer_dirty - prev->mark_buffer_dirty;
- w->cachestat.dirty = mbd;
+ w->cachestat.dirty = (long long)mbd;
uint64_t apcl = current->add_to_page_cache_lru - prev->add_to_page_cache_lru;
uint64_t apd = current->account_page_dirtied - prev->account_page_dirtied;
@@ -1029,6 +1053,7 @@ void ebpf_cache_send_apps_data(struct ebpf_target *root)
write_chart_dimension("misses", value);
ebpf_write_end_chart();
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -1073,7 +1098,7 @@ void ebpf_cachestat_calc_chart_values()
uint64_t mpa = current->mark_page_accessed - prev->mark_page_accessed;
uint64_t mbd = current->mark_buffer_dirty - prev->mark_buffer_dirty;
- ect->publish_cachestat.dirty = mbd;
+ ect->publish_cachestat.dirty = (long long)mbd;
uint64_t apcl = current->add_to_page_cache_lru - prev->add_to_page_cache_lru;
uint64_t apd = current->account_page_dirtied - prev->account_page_dirtied;
@@ -1090,35 +1115,82 @@ void ebpf_cachestat_calc_chart_values()
**/
static void ebpf_create_systemd_cachestat_charts(int update_every)
{
- ebpf_create_charts_on_systemd(NETDATA_CACHESTAT_HIT_RATIO_CHART,
- "Hit ratio",
- EBPF_COMMON_DIMENSION_PERCENTAGE, NETDATA_CACHESTAT_SUBMENU,
- NETDATA_EBPF_CHART_TYPE_LINE, 21100,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX],
- NETDATA_SYSTEMD_CACHESTAT_HIT_RATIO_CONTEXT, NETDATA_EBPF_MODULE_NAME_CACHESTAT,
- update_every);
+ static ebpf_systemd_args_t data_hit_ratio = {
+ .title = "Hit ratio",
+ .units = EBPF_COMMON_DIMENSION_PERCENTAGE,
+ .family = NETDATA_CACHESTAT_SUBMENU,
+ .charttype = NETDATA_EBPF_CHART_TYPE_LINE,
+ .order = 21100,
+ .algorithm = EBPF_CHART_ALGORITHM_ABSOLUTE,
+ .context = NETDATA_SYSTEMD_CACHESTAT_HIT_RATIO_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_CACHESTAT,
+ .update_every = 0,
+ .suffix = NETDATA_CACHESTAT_HIT_RATIO_CHART,
+ .dimension = "percentage"
+ };
- ebpf_create_charts_on_systemd(NETDATA_CACHESTAT_DIRTY_CHART,
- "Number of dirty pages",
- EBPF_CACHESTAT_DIMENSION_PAGE, NETDATA_CACHESTAT_SUBMENU,
- NETDATA_EBPF_CHART_TYPE_LINE, 21101,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX],
- NETDATA_SYSTEMD_CACHESTAT_MODIFIED_CACHE_CONTEXT, NETDATA_EBPF_MODULE_NAME_CACHESTAT,
- update_every);
+ static ebpf_systemd_args_t data_dirty = {
+ .title = "Number of dirty pages",
+ .units = EBPF_CACHESTAT_DIMENSION_PAGE,
+ .family = NETDATA_CACHESTAT_SUBMENU,
+ .charttype = NETDATA_EBPF_CHART_TYPE_LINE,
+ .order = 21101,
+ .algorithm = EBPF_CHART_ALGORITHM_ABSOLUTE,
+ .context = NETDATA_SYSTEMD_CACHESTAT_MODIFIED_CACHE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_CACHESTAT,
+ .update_every = 0,
+ .suffix = NETDATA_CACHESTAT_DIRTY_CHART,
+ .dimension = "pages"
+ };
- ebpf_create_charts_on_systemd(NETDATA_CACHESTAT_HIT_CHART, "Number of accessed files",
- EBPF_CACHESTAT_DIMENSION_HITS, NETDATA_CACHESTAT_SUBMENU,
- NETDATA_EBPF_CHART_TYPE_LINE, 21102,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX],
- NETDATA_SYSTEMD_CACHESTAT_HIT_FILE_CONTEXT, NETDATA_EBPF_MODULE_NAME_CACHESTAT,
- update_every);
+ static ebpf_systemd_args_t data_hit = {
+ .title = "Number of accessed pages",
+ .units = EBPF_CACHESTAT_DIMENSION_HITS,
+ .family = NETDATA_CACHESTAT_SUBMENU,
+ .charttype = NETDATA_EBPF_CHART_TYPE_LINE,
+ .order = 21102,
+ .algorithm = EBPF_CHART_ALGORITHM_ABSOLUTE,
+ .context = NETDATA_SYSTEMD_CACHESTAT_HIT_FILE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_CACHESTAT,
+ .update_every = 0,
+ .suffix = NETDATA_CACHESTAT_HIT_CHART,
+ .dimension = "hits"
+ };
- ebpf_create_charts_on_systemd(NETDATA_CACHESTAT_MISSES_CHART, "Files out of page cache",
- EBPF_CACHESTAT_DIMENSION_MISSES, NETDATA_CACHESTAT_SUBMENU,
- NETDATA_EBPF_CHART_TYPE_LINE, 21103,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX],
- NETDATA_SYSTEMD_CACHESTAT_MISS_FILES_CONTEXT, NETDATA_EBPF_MODULE_NAME_CACHESTAT,
- update_every);
+ static ebpf_systemd_args_t data_miss = {
+ .title = "Files out of page cache",
+ .units = EBPF_CACHESTAT_DIMENSION_MISSES,
+ .family = NETDATA_CACHESTAT_SUBMENU,
+ .charttype = NETDATA_EBPF_CHART_TYPE_LINE,
+ .order = 21103,
+ .algorithm = EBPF_CHART_ALGORITHM_ABSOLUTE,
+ .context = NETDATA_SYSTEMD_CACHESTAT_MISS_FILES_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_CACHESTAT,
+ .update_every = 0,
+ .suffix = NETDATA_CACHESTAT_MISSES_CHART,
+ .dimension = "misses"
+ };
+
+ if (!data_miss.update_every)
+ data_hit_ratio.update_every = data_dirty.update_every =
+ data_hit.update_every = data_miss.update_every = update_every;
+
+ ebpf_cgroup_target_t *w;
+ for (w = ebpf_cgroup_pids; w; w = w->next) {
+ if (unlikely(!w->systemd || w->flags & NETDATA_EBPF_SERVICES_HAS_CACHESTAT_CHART))
+ continue;
+
+ data_hit_ratio.id = data_dirty.id = data_hit.id = data_miss.id = w->name;
+ ebpf_create_charts_on_systemd(&data_hit_ratio);
+
+ ebpf_create_charts_on_systemd(&data_dirty);
+
+ ebpf_create_charts_on_systemd(&data_hit);
+
+ ebpf_create_charts_on_systemd(&data_miss);
+
+ w->flags |= NETDATA_EBPF_SERVICES_HAS_CACHESTAT_CHART;
+ }
}
/**
@@ -1130,37 +1202,27 @@ static void ebpf_send_systemd_cachestat_charts()
{
ebpf_cgroup_target_t *ect;
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_CACHESTAT_HIT_RATIO_CHART, "");
for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_cachestat.ratio);
+ if (unlikely(!(ect->flags & NETDATA_EBPF_SERVICES_HAS_CACHESTAT_CHART)) ) {
+ continue;
}
- }
- ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_CACHESTAT_DIRTY_CHART, "");
- for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_cachestat.dirty);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_CACHESTAT_HIT_RATIO_CHART);
+ write_chart_dimension("percentage", (long long)ect->publish_cachestat.ratio);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_CACHESTAT_HIT_CHART, "");
- for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_cachestat.hit);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_CACHESTAT_DIRTY_CHART);
+ write_chart_dimension("pages", (long long)ect->publish_cachestat.dirty);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_CACHESTAT_MISSES_CHART, "");
- for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_cachestat.miss);
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_CACHESTAT_HIT_CHART);
+ write_chart_dimension("hits", (long long)ect->publish_cachestat.hit);
+ ebpf_write_end_chart();
+
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_CACHESTAT_MISSES_CHART);
+ write_chart_dimension("misses", (long long)ect->publish_cachestat.miss);
+ ebpf_write_end_chart();
}
- ebpf_write_end_chart();
}
/**
@@ -1199,7 +1261,7 @@ static void ebpf_create_specific_cachestat_charts(char *type, int update_every)
{
ebpf_create_chart(type, NETDATA_CACHESTAT_HIT_RATIO_CHART,
"Hit ratio",
- EBPF_COMMON_DIMENSION_PERCENTAGE, NETDATA_CACHESTAT_CGROUP_SUBMENU,
+ EBPF_COMMON_DIMENSION_PERCENTAGE, NETDATA_CACHESTAT_SUBMENU,
NETDATA_CGROUP_CACHESTAT_HIT_RATIO_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5200,
ebpf_create_global_dimension,
@@ -1207,7 +1269,7 @@ static void ebpf_create_specific_cachestat_charts(char *type, int update_every)
ebpf_create_chart(type, NETDATA_CACHESTAT_DIRTY_CHART,
"Number of dirty pages",
- EBPF_CACHESTAT_DIMENSION_PAGE, NETDATA_CACHESTAT_CGROUP_SUBMENU,
+ EBPF_CACHESTAT_DIMENSION_PAGE, NETDATA_CACHESTAT_SUBMENU,
NETDATA_CGROUP_CACHESTAT_MODIFIED_CACHE_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5201,
ebpf_create_global_dimension,
@@ -1216,7 +1278,7 @@ static void ebpf_create_specific_cachestat_charts(char *type, int update_every)
ebpf_create_chart(type, NETDATA_CACHESTAT_HIT_CHART,
"Number of accessed files",
- EBPF_CACHESTAT_DIMENSION_HITS, NETDATA_CACHESTAT_CGROUP_SUBMENU,
+ EBPF_CACHESTAT_DIMENSION_HITS, NETDATA_CACHESTAT_SUBMENU,
NETDATA_CGROUP_CACHESTAT_HIT_FILES_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5202,
ebpf_create_global_dimension,
@@ -1225,7 +1287,7 @@ static void ebpf_create_specific_cachestat_charts(char *type, int update_every)
ebpf_create_chart(type, NETDATA_CACHESTAT_MISSES_CHART,
"Files out of page cache",
- EBPF_CACHESTAT_DIMENSION_MISSES, NETDATA_CACHESTAT_CGROUP_SUBMENU,
+ EBPF_CACHESTAT_DIMENSION_MISSES, NETDATA_CACHESTAT_SUBMENU,
NETDATA_CGROUP_CACHESTAT_MISS_FILES_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5203,
ebpf_create_global_dimension,
@@ -1279,15 +1341,11 @@ static void ebpf_obsolete_specific_cachestat_charts(char *type, int update_every
*/
void ebpf_cachestat_send_cgroup_data(int update_every)
{
- if (!ebpf_cgroup_pids)
- return;
-
pthread_mutex_lock(&mutex_cgroup_shm);
ebpf_cgroup_target_t *ect;
ebpf_cachestat_calc_chart_values();
- int has_systemd = shm_ebpf_cgroup.header->systemd_enabled;
- if (has_systemd) {
+ if (shm_ebpf_cgroup.header->systemd_enabled) {
if (send_cgroup_chart) {
ebpf_create_systemd_cachestat_charts(update_every);
}
@@ -1344,12 +1402,9 @@ static void cachestat_collector(ebpf_module_t *em)
counter = 0;
netdata_apps_integration_flags_t apps = em->apps_charts;
ebpf_cachestat_read_global_tables(stats, maps_per_core);
- pthread_mutex_lock(&collect_data_mutex);
- if (apps)
- ebpf_read_cachestat_apps_table(maps_per_core);
if (cgroups)
- ebpf_update_cachestat_cgroup(maps_per_core);
+ ebpf_update_cachestat_cgroup();
pthread_mutex_lock(&lock);
@@ -1358,16 +1413,10 @@ static void cachestat_collector(ebpf_module_t *em)
if (apps & NETDATA_EBPF_APPS_FLAG_CHART_CREATED)
ebpf_cache_send_apps_data(apps_groups_root_target);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_cachestat_pid)
- ebpf_send_data_aral_chart(ebpf_aral_cachestat_pid, em);
-#endif
-
if (cgroups)
ebpf_cachestat_send_cgroup_data(update_every);
pthread_mutex_unlock(&lock);
- pthread_mutex_unlock(&collect_data_mutex);
pthread_mutex_lock(&ebpf_exit_cleanup);
if (running_time && !em->running_time)
@@ -1442,17 +1491,10 @@ static void ebpf_create_memory_charts(ebpf_module_t *em)
*
* We are not testing the return, because callocz does this and shutdown the software
* case it was not possible to allocate.
- *
- * @param apps is apps enabled?
*/
-static void ebpf_cachestat_allocate_global_vectors(int apps)
+static void ebpf_cachestat_allocate_global_vectors()
{
- if (apps) {
- cachestat_pid = callocz((size_t)pid_max, sizeof(netdata_publish_cachestat_t *));
- ebpf_cachestat_aral_init();
- cachestat_vector = callocz((size_t)ebpf_nprocs, sizeof(netdata_cachestat_pid_t));
- }
-
+ cachestat_vector = callocz((size_t)ebpf_nprocs, sizeof(netdata_cachestat_pid_t));
cachestat_values = callocz((size_t)ebpf_nprocs, sizeof(netdata_idx_t));
memset(cachestat_hash_values, 0, NETDATA_CACHESTAT_END * sizeof(netdata_idx_t));
@@ -1560,7 +1602,7 @@ void *ebpf_cachestat_thread(void *ptr)
goto endcachestat;
}
- ebpf_cachestat_allocate_global_vectors(em->apps_charts);
+ ebpf_cachestat_allocate_global_vectors();
int algorithms[NETDATA_CACHESTAT_END] = {
NETDATA_EBPF_ABSOLUTE_IDX, NETDATA_EBPF_INCREMENTAL_IDX, NETDATA_EBPF_ABSOLUTE_IDX, NETDATA_EBPF_ABSOLUTE_IDX
@@ -1574,13 +1616,16 @@ void *ebpf_cachestat_thread(void *ptr)
ebpf_update_stats(&plugin_statistics, em);
ebpf_update_kernel_memory_with_vector(&plugin_statistics, em->maps, EBPF_ACTION_STAT_ADD);
ebpf_create_memory_charts(em);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_cachestat_pid)
- cachestat_disable_priority = ebpf_statistic_create_aral_chart(NETDATA_EBPF_CACHESTAT_ARAL_NAME, em);
-#endif
pthread_mutex_unlock(&lock);
+ ebpf_read_cachestat.thread = mallocz(sizeof(netdata_thread_t));
+ netdata_thread_create(ebpf_read_cachestat.thread,
+ ebpf_read_cachestat.name,
+ NETDATA_THREAD_OPTION_DEFAULT,
+ ebpf_read_cachestat_thread,
+ em);
+
cachestat_collector(em);
endcachestat:
diff --git a/collectors/ebpf.plugin/ebpf_cachestat.h b/src/collectors/ebpf.plugin/ebpf_cachestat.h
index ba2b12833..4578fbe98 100644
--- a/collectors/ebpf.plugin/ebpf_cachestat.h
+++ b/src/collectors/ebpf.plugin/ebpf_cachestat.h
@@ -14,7 +14,6 @@
#define NETDATA_CACHESTAT_MISSES_CHART "cachestat_misses"
#define NETDATA_CACHESTAT_SUBMENU "page_cache"
-#define NETDATA_CACHESTAT_CGROUP_SUBMENU "page cache (eBPF)"
#define EBPF_CACHESTAT_DIMENSION_PAGE "pages/s"
#define EBPF_CACHESTAT_DIMENSION_HITS "hits/s"
@@ -29,13 +28,10 @@
#define NETDATA_CGROUP_CACHESTAT_HIT_FILES_CONTEXT "cgroup.cachestat_hits"
#define NETDATA_CGROUP_CACHESTAT_MISS_FILES_CONTEXT "cgroup.cachestat_misses"
-#define NETDATA_SYSTEMD_CACHESTAT_HIT_RATIO_CONTEXT "services.cachestat_ratio"
-#define NETDATA_SYSTEMD_CACHESTAT_MODIFIED_CACHE_CONTEXT "services.cachestat_dirties"
-#define NETDATA_SYSTEMD_CACHESTAT_HIT_FILE_CONTEXT "services.cachestat_hits"
-#define NETDATA_SYSTEMD_CACHESTAT_MISS_FILES_CONTEXT "services.cachestat_misses"
-
-// ARAL Name
-#define NETDATA_EBPF_CACHESTAT_ARAL_NAME "ebpf_cachestat"
+#define NETDATA_SYSTEMD_CACHESTAT_HIT_RATIO_CONTEXT "systemd.services.cachestat_ratio"
+#define NETDATA_SYSTEMD_CACHESTAT_MODIFIED_CACHE_CONTEXT "systemd.services.cachestat_dirties"
+#define NETDATA_SYSTEMD_CACHESTAT_HIT_FILE_CONTEXT "systemd.services.cachestat_hits"
+#define NETDATA_SYSTEMD_CACHESTAT_MISS_FILES_CONTEXT "systemd.services.cachestat_misses"
// variables
enum cachestat_counters {
@@ -69,6 +65,12 @@ enum cachestat_tables {
};
typedef struct netdata_publish_cachestat_pid {
+ uint64_t ct;
+ uint32_t tgid;
+ uint32_t uid;
+ uint32_t gid;
+ char name[TASK_COMM_LEN];
+
uint64_t add_to_page_cache_lru;
uint64_t mark_page_accessed;
uint64_t account_page_dirtied;
@@ -76,6 +78,8 @@ typedef struct netdata_publish_cachestat_pid {
} netdata_cachestat_pid_t;
typedef struct netdata_publish_cachestat {
+ uint64_t ct;
+
long long ratio;
long long dirty;
long long hit;
diff --git a/collectors/ebpf.plugin/ebpf_cgroup.c b/src/collectors/ebpf.plugin/ebpf_cgroup.c
index 1aadfbaf8..881a00ddf 100644
--- a/collectors/ebpf.plugin/ebpf_cgroup.c
+++ b/src/collectors/ebpf.plugin/ebpf_cgroup.c
@@ -327,17 +327,22 @@ void ebpf_parse_cgroup_shm_data()
* @param module chart module name, this is the eBPF thread.
* @param update_every value to overwrite the update frequency set by the server.
*/
-void ebpf_create_charts_on_systemd(char *id, char *title, char *units, char *family, char *charttype, int order,
- char *algorithm, char *context, char *module, int update_every)
+void ebpf_create_charts_on_systemd(ebpf_systemd_args_t *chart)
{
- ebpf_cgroup_target_t *w;
- ebpf_write_chart_cmd(NETDATA_SERVICE_FAMILY, id, "", title, units, family, charttype, context,
- order, update_every, module);
-
- for (w = ebpf_cgroup_pids; w; w = w->next) {
- if (unlikely(w->systemd) && unlikely(w->updated))
- fprintf(stdout, "DIMENSION %s '' %s 1 1\n", w->name, algorithm);
- }
+ ebpf_write_chart_cmd(NETDATA_SERVICE_FAMILY,
+ chart->id,
+ chart->suffix,
+ chart->title,
+ chart->units,
+ chart->family,
+ chart->charttype,
+ chart->context,
+ chart->order,
+ chart->update_every,
+ chart->module);
+ ebpf_create_chart_labels("service_name", chart->id, 0);
+ ebpf_commit_label();
+ fprintf(stdout, "DIMENSION %s '' %s 1 1\n", chart->dimension, chart->algorithm);
}
// --------------------------------------------------------------------------------------------------------------------
diff --git a/collectors/ebpf.plugin/ebpf_dcstat.c b/src/collectors/ebpf.plugin/ebpf_dcstat.c
index 4ff6c82ab..7fa9ca1f0 100644
--- a/collectors/ebpf.plugin/ebpf_dcstat.c
+++ b/src/collectors/ebpf.plugin/ebpf_dcstat.c
@@ -59,9 +59,16 @@ netdata_ebpf_targets_t dc_targets[] = { {.name = "lookup_fast", .mode = EBPF_LOA
{.name = "d_lookup", .mode = EBPF_LOAD_TRAMPOLINE},
{.name = NULL, .mode = EBPF_LOAD_TRAMPOLINE}};
-#ifdef NETDATA_DEV_MODE
-int dcstat_disable_priority;
-#endif
+struct netdata_static_thread ebpf_read_dcstat = {
+ .name = "EBPF_READ_DCSTAT",
+ .config_section = NULL,
+ .config_name = NULL,
+ .env_name = NULL,
+ .enabled = 1,
+ .thread = NULL,
+ .init_routine = NULL,
+ .start_routine = NULL
+};
#ifdef LIBBPF_MAJOR_VERSION
/**
@@ -75,7 +82,6 @@ static inline void ebpf_dc_disable_probes(struct dc_bpf *obj)
{
bpf_program__set_autoload(obj->progs.netdata_lookup_fast_kprobe, false);
bpf_program__set_autoload(obj->progs.netdata_d_lookup_kretprobe, false);
- bpf_program__set_autoload(obj->progs.netdata_dcstat_release_task_kprobe, false);
}
/*
@@ -89,7 +95,6 @@ static inline void ebpf_dc_disable_trampoline(struct dc_bpf *obj)
{
bpf_program__set_autoload(obj->progs.netdata_lookup_fast_fentry, false);
bpf_program__set_autoload(obj->progs.netdata_d_lookup_fexit, false);
- bpf_program__set_autoload(obj->progs.netdata_dcstat_release_task_fentry, false);
}
/**
@@ -106,9 +111,6 @@ static void ebpf_dc_set_trampoline_target(struct dc_bpf *obj)
bpf_program__set_attach_target(obj->progs.netdata_d_lookup_fexit, 0,
dc_targets[NETDATA_DC_TARGET_D_LOOKUP].name);
-
- bpf_program__set_attach_target(obj->progs.netdata_dcstat_release_task_fentry, 0,
- EBPF_COMMON_FNCT_CLEAN_UP);
}
/**
@@ -125,7 +127,7 @@ static int ebpf_dc_attach_probes(struct dc_bpf *obj)
obj->links.netdata_d_lookup_kretprobe = bpf_program__attach_kprobe(obj->progs.netdata_d_lookup_kretprobe,
true,
dc_targets[NETDATA_DC_TARGET_D_LOOKUP].name);
- int ret = libbpf_get_error(obj->links.netdata_d_lookup_kretprobe);
+ long ret = libbpf_get_error(obj->links.netdata_d_lookup_kretprobe);
if (ret)
return -1;
@@ -140,13 +142,6 @@ static int ebpf_dc_attach_probes(struct dc_bpf *obj)
if (ret)
return -1;
- obj->links.netdata_dcstat_release_task_kprobe = bpf_program__attach_kprobe(obj->progs.netdata_dcstat_release_task_kprobe,
- false,
- EBPF_COMMON_FNCT_CLEAN_UP);
- ret = libbpf_get_error(obj->links.netdata_dcstat_release_task_kprobe);
- if (ret)
- return -1;
-
return 0;
}
@@ -206,19 +201,6 @@ netdata_ebpf_program_loaded_t ebpf_dc_update_load(ebpf_module_t *em)
}
/**
- * Disable Release Task
- *
- * Disable release task when apps is not enabled.
- *
- * @param obj is the main structure for bpf objects.
- */
-static void ebpf_dc_disable_release_task(struct dc_bpf *obj)
-{
- bpf_program__set_autoload(obj->progs.netdata_dcstat_release_task_kprobe, false);
- bpf_program__set_autoload(obj->progs.netdata_dcstat_release_task_fentry, false);
-}
-
-/**
* Load and attach
*
* Load and attach the eBPF code in kernel.
@@ -241,9 +223,6 @@ static inline int ebpf_dc_load_and_attach(struct dc_bpf *obj, ebpf_module_t *em)
ebpf_dc_adjust_map(obj, em);
- if (!em->apps_charts && !em->cgroup_charts)
- ebpf_dc_disable_release_task(obj);
-
int ret = dc_bpf__load(obj);
if (ret) {
return ret;
@@ -298,11 +277,11 @@ static void ebpf_obsolete_specific_dc_charts(char *type, int update_every);
*
* @param em a pointer to `struct ebpf_module`
*/
-static void ebpf_obsolete_dc_services(ebpf_module_t *em)
+static void ebpf_obsolete_dc_services(ebpf_module_t *em, char *id)
{
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_DC_HIT_CHART,
- "",
"Percentage of files inside directory cache",
EBPF_COMMON_DIMENSION_PERCENTAGE,
NETDATA_DIRECTORY_CACHE_SUBMENU,
@@ -312,8 +291,8 @@ static void ebpf_obsolete_dc_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_DC_REFERENCE_CHART,
- "",
"Count file access",
EBPF_COMMON_DIMENSION_FILES,
NETDATA_DIRECTORY_CACHE_SUBMENU,
@@ -323,8 +302,8 @@ static void ebpf_obsolete_dc_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_DC_REQUEST_NOT_CACHE_CHART,
- "",
"Files not present inside directory cache",
EBPF_COMMON_DIMENSION_FILES,
NETDATA_DIRECTORY_CACHE_SUBMENU,
@@ -334,14 +313,14 @@ static void ebpf_obsolete_dc_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_DC_REQUEST_NOT_FOUND_CHART,
- "",
"Files not found",
EBPF_COMMON_DIMENSION_FILES,
NETDATA_DIRECTORY_CACHE_SUBMENU,
NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_SYSTEMD_DC_NOT_FOUND_CONTEXT,
- 21202,
+ 21203,
em->update_every);
}
@@ -355,12 +334,13 @@ static void ebpf_obsolete_dc_services(ebpf_module_t *em)
static inline void ebpf_obsolete_dc_cgroup_charts(ebpf_module_t *em) {
pthread_mutex_lock(&mutex_cgroup_shm);
- ebpf_obsolete_dc_services(em);
-
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (ect->systemd)
+ if (ect->systemd) {
+ ebpf_obsolete_dc_services(em, ect->name);
+
continue;
+ }
ebpf_obsolete_specific_dc_charts(ect->name, em->update_every);
}
@@ -378,6 +358,7 @@ void ebpf_obsolete_dc_apps_charts(struct ebpf_module *em)
{
struct ebpf_target *w;
int update_every = em->update_every;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = apps_groups_root_target; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_DCSTAT_IDX))))
continue;
@@ -428,6 +409,7 @@ void ebpf_obsolete_dc_apps_charts(struct ebpf_module *em)
w->charts_created &= ~(1<<EBPF_MODULE_DCSTAT_IDX);
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -473,6 +455,9 @@ static void ebpf_dcstat_exit(void *ptr)
{
ebpf_module_t *em = (ebpf_module_t *)ptr;
+ if (ebpf_read_dcstat.thread)
+ netdata_thread_cancel(*ebpf_read_dcstat.thread);
+
if (em->enabled == NETDATA_THREAD_EBPF_FUNCTION_RUNNING) {
pthread_mutex_lock(&lock);
if (em->cgroup_charts) {
@@ -486,11 +471,6 @@ static void ebpf_dcstat_exit(void *ptr)
ebpf_obsolete_dc_global(em);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_dcstat_pid)
- ebpf_statistic_obsolete_aral_chart(em, dcstat_disable_priority);
-#endif
-
fflush(stdout);
pthread_mutex_unlock(&lock);
}
@@ -523,6 +503,173 @@ static void ebpf_dcstat_exit(void *ptr)
*****************************************************************/
/**
+ * Apps Accumulator
+ *
+ * Sum all values read from kernel and store in the first address.
+ *
+ * @param out the vector with read values.
+ * @param maps_per_core do I need to read all cores?
+ */
+static void ebpf_dcstat_apps_accumulator(netdata_dcstat_pid_t *out, int maps_per_core)
+{
+ int i, end = (maps_per_core) ? ebpf_nprocs : 1;
+ netdata_dcstat_pid_t *total = &out[0];
+ uint64_t ct = total->ct;
+ for (i = 1; i < end; i++) {
+ netdata_dcstat_pid_t *w = &out[i];
+ total->cache_access += w->cache_access;
+ total->file_system += w->file_system;
+ total->not_found += w->not_found;
+
+ if (w->ct > ct)
+ ct = w->ct;
+ }
+ total->ct = ct;
+}
+
+/**
+ * Read Directory Cache APPS table
+ *
+ * Read the apps table and store data inside the structure.
+ *
+ * @param maps_per_core do I need to read all cores?
+ */
+static void ebpf_read_dc_apps_table(int maps_per_core, int max_period)
+{
+ netdata_dcstat_pid_t *cv = dcstat_vector;
+ int fd = dcstat_maps[NETDATA_DCSTAT_PID_STATS].map_fd;
+ size_t length = sizeof(netdata_dcstat_pid_t);
+ if (maps_per_core)
+ length *= ebpf_nprocs;
+
+ uint32_t key = 0, next_key = 0;
+ while (bpf_map_get_next_key(fd, &key, &next_key) == 0) {
+ if (bpf_map_lookup_elem(fd, &key, cv)) {
+ goto end_dc_loop;
+ }
+
+ ebpf_dcstat_apps_accumulator(cv, maps_per_core);
+
+ ebpf_pid_stat_t *pid_stat = ebpf_get_pid_entry(key, cv->tgid);
+ if (pid_stat) {
+ netdata_publish_dcstat_t *publish = &pid_stat->dc;
+ if (!publish->ct || publish->ct != cv->ct) {
+ memcpy(&publish->curr, &cv[0], sizeof(netdata_dcstat_pid_t));
+ pid_stat->not_updated = 0;
+ } else if (++pid_stat->not_updated >= max_period) {
+ bpf_map_delete_elem(fd, &key);
+ pid_stat->not_updated = 0;
+ }
+ }
+
+end_dc_loop:
+ // We are cleaning to avoid passing data read from one process to other.
+ memset(cv, 0, length);
+ key = next_key;
+ }
+}
+
+/**
+ * Cachestat sum PIDs
+ *
+ * Sum values for all PIDs associated to a group
+ *
+ * @param publish output structure.
+ * @param root structure with listed IPs
+ */
+void ebpf_dcstat_sum_pids(netdata_publish_dcstat_t *publish, struct ebpf_pid_on_target *root)
+{
+ memset(&publish->curr, 0, sizeof(netdata_dcstat_pid_t));
+ netdata_dcstat_pid_t *dst = &publish->curr;
+ while (root) {
+ int32_t pid = root->pid;
+ ebpf_pid_stat_t *pid_stat = ebpf_get_pid_entry(pid, 0);
+ if (pid_stat) {
+ netdata_publish_dcstat_t *w = &pid_stat->dc;
+ netdata_dcstat_pid_t *src = &w->curr;
+ dst->cache_access += src->cache_access;
+ dst->file_system += src->file_system;
+ dst->not_found += src->not_found;
+ }
+
+ root = root->next;
+ }
+}
+
+/**
+ * Resume apps data
+ */
+void ebpf_dc_resume_apps_data()
+{
+ struct ebpf_target *w;
+
+ for (w = apps_groups_root_target; w; w = w->next) {
+ if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_DCSTAT_IDX))))
+ continue;
+
+ ebpf_dcstat_sum_pids(&w->dcstat, w->root_pid);
+
+ uint64_t cache = w->dcstat.curr.cache_access;
+ uint64_t not_found = w->dcstat.curr.not_found;
+
+ dcstat_update_publish(&w->dcstat, cache, not_found);
+ }
+}
+
+/**
+ * DCstat thread
+ *
+ * Thread used to generate dcstat charts.
+ *
+ * @param ptr a pointer to `struct ebpf_module`
+ *
+ * @return It always return NULL
+ */
+void *ebpf_read_dcstat_thread(void *ptr)
+{
+ heartbeat_t hb;
+ heartbeat_init(&hb);
+
+ ebpf_module_t *em = (ebpf_module_t *)ptr;
+
+ int maps_per_core = em->maps_per_core;
+ int update_every = em->update_every;
+
+ int counter = update_every - 1;
+
+ uint32_t lifetime = em->lifetime;
+ uint32_t running_time = 0;
+ usec_t period = update_every * USEC_PER_SEC;
+ int max_period = update_every * EBPF_CLEANUP_FACTOR;
+ while (!ebpf_plugin_exit && running_time < lifetime) {
+ (void)heartbeat_next(&hb, period);
+ if (ebpf_plugin_exit || ++counter != update_every)
+ continue;
+
+ netdata_thread_disable_cancelability();
+
+ pthread_mutex_lock(&collect_data_mutex);
+ ebpf_read_dc_apps_table(maps_per_core, max_period);
+ ebpf_dc_resume_apps_data();
+ pthread_mutex_unlock(&collect_data_mutex);
+
+ counter = 0;
+
+ pthread_mutex_lock(&ebpf_exit_cleanup);
+ if (running_time && !em->running_time)
+ running_time = update_every;
+ else
+ running_time += update_every;
+
+ em->running_time = running_time;
+ pthread_mutex_unlock(&ebpf_exit_cleanup);
+ netdata_thread_enable_cancelability();
+ }
+
+ return NULL;
+}
+
+/**
* Create apps charts
*
* Call ebpf_create_chart to create the charts on apps submenu.
@@ -611,106 +758,14 @@ void ebpf_dcstat_create_apps_charts(struct ebpf_module *em, void *ptr)
*****************************************************************/
/**
- * Apps Accumulator
- *
- * Sum all values read from kernel and store in the first address.
- *
- * @param out the vector with read values.
- * @param maps_per_core do I need to read all cores?
- */
-static void dcstat_apps_accumulator(netdata_dcstat_pid_t *out, int maps_per_core)
-{
- int i, end = (maps_per_core) ? ebpf_nprocs : 1;
- netdata_dcstat_pid_t *total = &out[0];
- for (i = 1; i < end; i++) {
- netdata_dcstat_pid_t *w = &out[i];
- total->cache_access += w->cache_access;
- total->file_system += w->file_system;
- total->not_found += w->not_found;
- }
-}
-
-/**
- * Save PID values
- *
- * Save the current values inside the structure
- *
- * @param out vector used to plot charts
- * @param publish vector with values read from hash tables.
- */
-static inline void dcstat_save_pid_values(netdata_publish_dcstat_t *out, netdata_dcstat_pid_t *publish)
-{
- memcpy(&out->curr, &publish[0], sizeof(netdata_dcstat_pid_t));
-}
-
-/**
- * Fill PID
- *
- * Fill PID structures
- *
- * @param current_pid pid that we are collecting data
- * @param out values read from hash tables;
- */
-static void dcstat_fill_pid(uint32_t current_pid, netdata_dcstat_pid_t *publish)
-{
- netdata_publish_dcstat_t *curr = dcstat_pid[current_pid];
- if (!curr) {
- curr = ebpf_publish_dcstat_get();
- dcstat_pid[current_pid] = curr;
- }
-
- dcstat_save_pid_values(curr, publish);
-}
-
-/**
- * Read Directory Cache APPS table
- *
- * Read the apps table and store data inside the structure.
- *
- * @param maps_per_core do I need to read all cores?
- */
-static void read_dc_apps_table(int maps_per_core)
-{
- netdata_dcstat_pid_t *cv = dcstat_vector;
- uint32_t key;
- struct ebpf_pid_stat *pids = ebpf_root_of_pids;
- int fd = dcstat_maps[NETDATA_DCSTAT_PID_STATS].map_fd;
- size_t length = sizeof(netdata_dcstat_pid_t);
- if (maps_per_core)
- length *= ebpf_nprocs;
-
- while (pids) {
- key = pids->pid;
-
- if (bpf_map_lookup_elem(fd, &key, cv)) {
- pids = pids->next;
- continue;
- }
-
- dcstat_apps_accumulator(cv, maps_per_core);
-
- dcstat_fill_pid(key, cv);
-
- // We are cleaning to avoid passing data read from one process to other.
- memset(cv, 0, length);
-
- pids = pids->next;
- }
-}
-
-/**
* Update cgroup
*
* Update cgroup data based in collected PID.
*
* @param maps_per_core do I need to read all cores?
*/
-static void ebpf_update_dc_cgroup(int maps_per_core)
+static void ebpf_update_dc_cgroup()
{
- netdata_dcstat_pid_t *cv = dcstat_vector;
- int fd = dcstat_maps[NETDATA_DCSTAT_PID_STATS].map_fd;
- size_t length = sizeof(netdata_dcstat_pid_t)*ebpf_nprocs;
-
ebpf_cgroup_target_t *ect;
pthread_mutex_lock(&mutex_cgroup_shm);
for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
@@ -718,19 +773,11 @@ static void ebpf_update_dc_cgroup(int maps_per_core)
for (pids = ect->pids; pids; pids = pids->next) {
int pid = pids->pid;
netdata_dcstat_pid_t *out = &pids->dc;
- if (likely(dcstat_pid) && dcstat_pid[pid]) {
- netdata_publish_dcstat_t *in = dcstat_pid[pid];
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ netdata_publish_dcstat_t *in = &local_pid->dc;
memcpy(out, &in->curr, sizeof(netdata_dcstat_pid_t));
- } else {
- memset(cv, 0, length);
- if (bpf_map_lookup_elem(fd, &pid, cv)) {
- continue;
- }
-
- dcstat_apps_accumulator(cv, maps_per_core);
-
- memcpy(out, cv, sizeof(netdata_dcstat_pid_t));
}
}
}
@@ -763,32 +810,6 @@ static void ebpf_dc_read_global_tables(netdata_idx_t *stats, int maps_per_core)
}
/**
- * Cachestat sum PIDs
- *
- * Sum values for all PIDs associated to a group
- *
- * @param publish output structure.
- * @param root structure with listed IPs
- */
-void ebpf_dcstat_sum_pids(netdata_publish_dcstat_t *publish, struct ebpf_pid_on_target *root)
-{
- memset(&publish->curr, 0, sizeof(netdata_dcstat_pid_t));
- netdata_dcstat_pid_t *dst = &publish->curr;
- while (root) {
- int32_t pid = root->pid;
- netdata_publish_dcstat_t *w = dcstat_pid[pid];
- if (w) {
- netdata_dcstat_pid_t *src = &w->curr;
- dst->cache_access += src->cache_access;
- dst->file_system += src->file_system;
- dst->not_found += src->not_found;
- }
-
- root = root->next;
- }
-}
-
-/**
* Send data to Netdata calling auxiliary functions.
*
* @param root the target list.
@@ -798,17 +819,11 @@ void ebpf_dcache_send_apps_data(struct ebpf_target *root)
struct ebpf_target *w;
collected_number value;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = root; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_DCSTAT_IDX))))
continue;
- ebpf_dcstat_sum_pids(&w->dcstat, w->root_pid);
-
- uint64_t cache = w->dcstat.curr.cache_access;
- uint64_t not_found = w->dcstat.curr.not_found;
-
- dcstat_update_publish(&w->dcstat, cache, not_found);
-
value = (collected_number) w->dcstat.ratio;
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_dc_hit");
write_chart_dimension("ratio", value);
@@ -845,6 +860,7 @@ void ebpf_dcache_send_apps_data(struct ebpf_target *root)
ebpf_write_end_chart();
w->dcstat.prev.not_found = w->dcstat.curr.not_found;
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -1024,45 +1040,82 @@ void ebpf_dc_calc_chart_values()
**/
static void ebpf_create_systemd_dc_charts(int update_every)
{
- ebpf_create_charts_on_systemd(NETDATA_DC_HIT_CHART,
- "Percentage of files inside directory cache",
- EBPF_COMMON_DIMENSION_PERCENTAGE,
- NETDATA_DIRECTORY_CACHE_SUBMENU,
- NETDATA_EBPF_CHART_TYPE_LINE,
- 21200,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX],
- NETDATA_SYSTEMD_DC_HIT_RATIO_CONTEXT, NETDATA_EBPF_MODULE_NAME_DCSTAT,
- update_every);
+ static ebpf_systemd_args_t data_dc_hit_ratio = {
+ .title = "Percentage of files inside directory cache",
+ .units = EBPF_COMMON_DIMENSION_PERCENTAGE,
+ .family = NETDATA_DIRECTORY_CACHE_SUBMENU,
+ .charttype = NETDATA_EBPF_CHART_TYPE_LINE,
+ .order = 21200,
+ .algorithm = EBPF_CHART_ALGORITHM_ABSOLUTE,
+ .context = NETDATA_SYSTEMD_DC_HIT_RATIO_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_DCSTAT,
+ .update_every = 0,
+ .suffix = NETDATA_DC_HIT_CHART,
+ .dimension = "percentage"
+ };
- ebpf_create_charts_on_systemd(NETDATA_DC_REFERENCE_CHART,
- "Count file access",
- EBPF_COMMON_DIMENSION_FILES,
- NETDATA_DIRECTORY_CACHE_SUBMENU,
- NETDATA_EBPF_CHART_TYPE_LINE,
- 21201,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX],
- NETDATA_SYSTEMD_DC_REFERENCE_CONTEXT, NETDATA_EBPF_MODULE_NAME_DCSTAT,
- update_every);
+ static ebpf_systemd_args_t data_dc_references = {
+ .title = "Count file access",
+ .units = EBPF_COMMON_DIMENSION_FILES,
+ .family = NETDATA_DIRECTORY_CACHE_SUBMENU,
+ .charttype = NETDATA_EBPF_CHART_TYPE_LINE,
+ .order = 21201,
+ .algorithm = EBPF_CHART_ALGORITHM_ABSOLUTE,
+ .context = NETDATA_SYSTEMD_DC_REFERENCE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_DCSTAT,
+ .update_every = 0,
+ .suffix = NETDATA_DC_REFERENCE_CHART,
+ .dimension = "files"
+ };
- ebpf_create_charts_on_systemd(NETDATA_DC_REQUEST_NOT_CACHE_CHART,
- "Files not present inside directory cache",
- EBPF_COMMON_DIMENSION_FILES,
- NETDATA_DIRECTORY_CACHE_SUBMENU,
- NETDATA_EBPF_CHART_TYPE_LINE,
- 21202,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX],
- NETDATA_SYSTEMD_DC_NOT_CACHE_CONTEXT, NETDATA_EBPF_MODULE_NAME_DCSTAT,
- update_every);
+ static ebpf_systemd_args_t data_dc_not_cache = {
+ .title = "Files not present inside directory cache",
+ .units = EBPF_COMMON_DIMENSION_FILES,
+ .family = NETDATA_DIRECTORY_CACHE_SUBMENU,
+ .charttype = NETDATA_EBPF_CHART_TYPE_LINE,
+ .order = 21202,
+ .algorithm = EBPF_CHART_ALGORITHM_ABSOLUTE,
+ .context = NETDATA_SYSTEMD_DC_NOT_CACHE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_DCSTAT,
+ .update_every = 0,
+ .suffix = NETDATA_DC_REQUEST_NOT_CACHE_CHART,
+ .dimension = "files"
+ };
- ebpf_create_charts_on_systemd(NETDATA_DC_REQUEST_NOT_FOUND_CHART,
- "Files not found",
- EBPF_COMMON_DIMENSION_FILES,
- NETDATA_DIRECTORY_CACHE_SUBMENU,
- NETDATA_EBPF_CHART_TYPE_LINE,
- 21202,
- ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX],
- NETDATA_SYSTEMD_DC_NOT_FOUND_CONTEXT, NETDATA_EBPF_MODULE_NAME_DCSTAT,
- update_every);
+ static ebpf_systemd_args_t data_dc_not_found = {
+ .title = "Files not found",
+ .units = EBPF_COMMON_DIMENSION_FILES,
+ .family = NETDATA_DIRECTORY_CACHE_SUBMENU,
+ .charttype = NETDATA_EBPF_CHART_TYPE_LINE,
+ .order = 21203,
+ .algorithm = EBPF_CHART_ALGORITHM_ABSOLUTE,
+ .context = NETDATA_SYSTEMD_DC_NOT_CACHE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_DCSTAT,
+ .update_every = 0,
+ .suffix = NETDATA_DC_REQUEST_NOT_FOUND_CHART,
+ .dimension = "files"
+ };
+
+ if (!data_dc_not_cache.update_every)
+ data_dc_hit_ratio.update_every = data_dc_not_cache.update_every =
+ data_dc_not_found.update_every = data_dc_references.update_every = update_every;
+
+ ebpf_cgroup_target_t *w;
+ for (w = ebpf_cgroup_pids; w; w = w->next) {
+ if (unlikely(!w->systemd || w->flags & NETDATA_EBPF_SERVICES_HAS_DC_CHART))
+ continue;
+
+ data_dc_hit_ratio.id = data_dc_not_cache.id = data_dc_not_found.id = data_dc_references.id = w->name;
+ ebpf_create_charts_on_systemd(&data_dc_hit_ratio);
+
+ ebpf_create_charts_on_systemd(&data_dc_not_found);
+
+ ebpf_create_charts_on_systemd(&data_dc_not_cache);
+
+ ebpf_create_charts_on_systemd(&data_dc_references);
+
+ w->flags |= NETDATA_EBPF_SERVICES_HAS_DC_CHART;
+ }
}
/**
@@ -1072,48 +1125,37 @@ static void ebpf_create_systemd_dc_charts(int update_every)
*/
static void ebpf_send_systemd_dc_charts()
{
- collected_number value;
ebpf_cgroup_target_t *ect;
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_DC_HIT_CHART, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long) ect->publish_dc.ratio);
+ collected_number value;
+ for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
+ if (unlikely(!(ect->flags & NETDATA_EBPF_SERVICES_HAS_DC_CHART)) ) {
+ continue;
}
- }
- ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_DC_REFERENCE_CHART, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long) ect->publish_dc.cache_access);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_DC_HIT_CHART);
+ write_chart_dimension("percentage", (long long) ect->publish_dc.ratio);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_DC_REQUEST_NOT_CACHE_CHART, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- value = (collected_number) (!ect->publish_dc.cache_access) ? 0 :
- (long long )ect->publish_dc.curr.file_system - (long long)ect->publish_dc.prev.file_system;
- ect->publish_dc.prev.file_system = ect->publish_dc.curr.file_system;
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_DC_REFERENCE_CHART);
+ write_chart_dimension("files", (long long) ect->publish_dc.cache_access);
+ ebpf_write_end_chart();
- write_chart_dimension(ect->name, (long long) value);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_DC_REQUEST_NOT_CACHE_CHART);
+ value = (collected_number) (!ect->publish_dc.cache_access) ? 0 :
+ (long long )ect->publish_dc.curr.file_system - (long long)ect->publish_dc.prev.file_system;
+ ect->publish_dc.prev.file_system = ect->publish_dc.curr.file_system;
+ write_chart_dimension("files", (long long) value);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_DC_REQUEST_NOT_FOUND_CHART, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- value = (collected_number) (!ect->publish_dc.cache_access) ? 0 :
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_DC_REQUEST_NOT_FOUND_CHART);
+ value = (collected_number) (!ect->publish_dc.cache_access) ? 0 :
(long long)ect->publish_dc.curr.not_found - (long long)ect->publish_dc.prev.not_found;
- ect->publish_dc.prev.not_found = ect->publish_dc.curr.not_found;
+ ect->publish_dc.prev.not_found = ect->publish_dc.curr.not_found;
- write_chart_dimension(ect->name, (long long) value);
- }
+ write_chart_dimension("files", (long long) value);
+ ebpf_write_end_chart();
}
- ebpf_write_end_chart();
}
/**
@@ -1159,15 +1201,11 @@ static void ebpf_send_specific_dc_data(char *type, netdata_publish_dcstat_t *pdc
*/
void ebpf_dc_send_cgroup_data(int update_every)
{
- if (!ebpf_cgroup_pids)
- return;
-
pthread_mutex_lock(&mutex_cgroup_shm);
ebpf_cgroup_target_t *ect;
ebpf_dc_calc_chart_values();
- int has_systemd = shm_ebpf_cgroup.header->systemd_enabled;
- if (has_systemd) {
+ if (shm_ebpf_cgroup.header->systemd_enabled) {
if (send_cgroup_chart) {
ebpf_create_systemd_dc_charts(update_every);
}
@@ -1223,12 +1261,9 @@ static void dcstat_collector(ebpf_module_t *em)
counter = 0;
netdata_apps_integration_flags_t apps = em->apps_charts;
ebpf_dc_read_global_tables(stats, maps_per_core);
- pthread_mutex_lock(&collect_data_mutex);
- if (apps)
- read_dc_apps_table(maps_per_core);
if (cgroups)
- ebpf_update_dc_cgroup(maps_per_core);
+ ebpf_update_dc_cgroup();
pthread_mutex_lock(&lock);
@@ -1237,16 +1272,10 @@ static void dcstat_collector(ebpf_module_t *em)
if (apps & NETDATA_EBPF_APPS_FLAG_CHART_CREATED)
ebpf_dcache_send_apps_data(apps_groups_root_target);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_dcstat_pid)
- ebpf_send_data_aral_chart(ebpf_aral_dcstat_pid, em);
-#endif
-
if (cgroups)
ebpf_dc_send_cgroup_data(update_every);
pthread_mutex_unlock(&lock);
- pthread_mutex_unlock(&collect_data_mutex);
pthread_mutex_lock(&ebpf_exit_cleanup);
if (running_time && !em->running_time)
@@ -1301,17 +1330,10 @@ static void ebpf_create_dc_global_charts(int update_every)
*
* We are not testing the return, because callocz does this and shutdown the software
* case it was not possible to allocate.
- *
- * @param apps is apps enabled?
*/
-static void ebpf_dcstat_allocate_global_vectors(int apps)
+static void ebpf_dcstat_allocate_global_vectors()
{
- if (apps) {
- ebpf_dcstat_aral_init();
- dcstat_pid = callocz((size_t)pid_max, sizeof(netdata_publish_dcstat_t *));
- dcstat_vector = callocz((size_t)ebpf_nprocs, sizeof(netdata_dcstat_pid_t));
- }
-
+ dcstat_vector = callocz((size_t)ebpf_nprocs, sizeof(netdata_dcstat_pid_t));
dcstat_values = callocz((size_t)ebpf_nprocs, sizeof(netdata_idx_t));
memset(dcstat_counter_aggregated_data, 0, NETDATA_DCSTAT_IDX_END * sizeof(netdata_syscall_stat_t));
@@ -1388,7 +1410,7 @@ void *ebpf_dcstat_thread(void *ptr)
goto enddcstat;
}
- ebpf_dcstat_allocate_global_vectors(em->apps_charts);
+ ebpf_dcstat_allocate_global_vectors();
int algorithms[NETDATA_DCSTAT_IDX_END] = {
NETDATA_EBPF_ABSOLUTE_IDX, NETDATA_EBPF_ABSOLUTE_IDX, NETDATA_EBPF_ABSOLUTE_IDX,
@@ -1403,13 +1425,16 @@ void *ebpf_dcstat_thread(void *ptr)
ebpf_create_dc_global_charts(em->update_every);
ebpf_update_stats(&plugin_statistics, em);
ebpf_update_kernel_memory_with_vector(&plugin_statistics, em->maps, EBPF_ACTION_STAT_ADD);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_dcstat_pid)
- dcstat_disable_priority = ebpf_statistic_create_aral_chart(NETDATA_EBPF_DCSTAT_ARAL_NAME, em);
-#endif
pthread_mutex_unlock(&lock);
+ ebpf_read_dcstat.thread = mallocz(sizeof(netdata_thread_t));
+ netdata_thread_create(ebpf_read_dcstat.thread,
+ ebpf_read_dcstat.name,
+ NETDATA_THREAD_OPTION_DEFAULT,
+ ebpf_read_dcstat_thread,
+ em);
+
dcstat_collector(em);
enddcstat:
diff --git a/collectors/ebpf.plugin/ebpf_dcstat.h b/src/collectors/ebpf.plugin/ebpf_dcstat.h
index 4d6aff12e..68f6d6131 100644
--- a/collectors/ebpf.plugin/ebpf_dcstat.h
+++ b/src/collectors/ebpf.plugin/ebpf_dcstat.h
@@ -24,10 +24,10 @@
#define NETDATA_CGROUP_DC_NOT_CACHE_CONTEXT "cgroup.dc_not_cache"
#define NETDATA_CGROUP_DC_NOT_FOUND_CONTEXT "cgroup.dc_not_found"
-#define NETDATA_SYSTEMD_DC_HIT_RATIO_CONTEXT "services.dc_ratio"
-#define NETDATA_SYSTEMD_DC_REFERENCE_CONTEXT "services.dc_reference"
-#define NETDATA_SYSTEMD_DC_NOT_CACHE_CONTEXT "services.dc_not_cache"
-#define NETDATA_SYSTEMD_DC_NOT_FOUND_CONTEXT "services.dc_not_found"
+#define NETDATA_SYSTEMD_DC_HIT_RATIO_CONTEXT "systemd.services.dc_ratio"
+#define NETDATA_SYSTEMD_DC_REFERENCE_CONTEXT "systemd.services.dc_reference"
+#define NETDATA_SYSTEMD_DC_NOT_CACHE_CONTEXT "systemd.services.dc_not_cache"
+#define NETDATA_SYSTEMD_DC_NOT_FOUND_CONTEXT "systemd.services.dc_not_found"
// ARAL name
#define NETDATA_EBPF_DCSTAT_ARAL_NAME "ebpf_dcstat"
@@ -64,12 +64,20 @@ enum directory_cache_targets {
};
typedef struct netdata_publish_dcstat_pid {
+ uint64_t ct;
+ uint32_t tgid;
+ uint32_t uid;
+ uint32_t gid;
+ char name[TASK_COMM_LEN];
+
uint64_t cache_access;
uint64_t file_system;
uint64_t not_found;
} netdata_dcstat_pid_t;
typedef struct netdata_publish_dcstat {
+ uint64_t ct;
+
long long ratio;
long long cache_access;
diff --git a/collectors/ebpf.plugin/ebpf_disk.c b/src/collectors/ebpf.plugin/ebpf_disk.c
index 466c2e3bb..466c2e3bb 100644
--- a/collectors/ebpf.plugin/ebpf_disk.c
+++ b/src/collectors/ebpf.plugin/ebpf_disk.c
diff --git a/collectors/ebpf.plugin/ebpf_disk.h b/src/collectors/ebpf.plugin/ebpf_disk.h
index 487ed376d..487ed376d 100644
--- a/collectors/ebpf.plugin/ebpf_disk.h
+++ b/src/collectors/ebpf.plugin/ebpf_disk.h
diff --git a/collectors/ebpf.plugin/ebpf_fd.c b/src/collectors/ebpf.plugin/ebpf_fd.c
index 3c8f30d3e..e1dc5b356 100644
--- a/collectors/ebpf.plugin/ebpf_fd.c
+++ b/src/collectors/ebpf.plugin/ebpf_fd.c
@@ -57,9 +57,16 @@ netdata_ebpf_targets_t fd_targets[] = { {.name = "open", .mode = EBPF_LOAD_TRAMP
{.name = "close", .mode = EBPF_LOAD_TRAMPOLINE},
{.name = NULL, .mode = EBPF_LOAD_TRAMPOLINE}};
-#ifdef NETDATA_DEV_MODE
-int fd_disable_priority;
-#endif
+struct netdata_static_thread ebpf_read_fd = {
+ .name = "EBPF_READ_FD",
+ .config_section = NULL,
+ .config_name = NULL,
+ .env_name = NULL,
+ .enabled = 1,
+ .thread = NULL,
+ .init_routine = NULL,
+ .start_routine = NULL
+};
#ifdef LIBBPF_MAJOR_VERSION
/**
@@ -73,7 +80,6 @@ static inline void ebpf_fd_disable_probes(struct fd_bpf *obj)
{
bpf_program__set_autoload(obj->progs.netdata_sys_open_kprobe, false);
bpf_program__set_autoload(obj->progs.netdata_sys_open_kretprobe, false);
- bpf_program__set_autoload(obj->progs.netdata_release_task_fd_kprobe, false);
if (!strcmp(fd_targets[NETDATA_FD_SYSCALL_CLOSE].name, close_targets[NETDATA_FD_CLOSE_FD])) {
bpf_program__set_autoload(obj->progs.netdata___close_fd_kretprobe, false);
bpf_program__set_autoload(obj->progs.netdata___close_fd_kprobe, false);
@@ -118,7 +124,6 @@ static inline void ebpf_disable_trampoline(struct fd_bpf *obj)
bpf_program__set_autoload(obj->progs.netdata_close_fd_fexit, false);
bpf_program__set_autoload(obj->progs.netdata___close_fd_fentry, false);
bpf_program__set_autoload(obj->progs.netdata___close_fd_fexit, false);
- bpf_program__set_autoload(obj->progs.netdata_release_task_fd_fentry, false);
}
/*
@@ -150,7 +155,6 @@ static void ebpf_set_trampoline_target(struct fd_bpf *obj)
{
bpf_program__set_attach_target(obj->progs.netdata_sys_open_fentry, 0, fd_targets[NETDATA_FD_SYSCALL_OPEN].name);
bpf_program__set_attach_target(obj->progs.netdata_sys_open_fexit, 0, fd_targets[NETDATA_FD_SYSCALL_OPEN].name);
- bpf_program__set_attach_target(obj->progs.netdata_release_task_fd_fentry, 0, EBPF_COMMON_FNCT_CLEAN_UP);
if (!strcmp(fd_targets[NETDATA_FD_SYSCALL_CLOSE].name, close_targets[NETDATA_FD_CLOSE_FD])) {
bpf_program__set_attach_target(
@@ -177,7 +181,7 @@ static int ebpf_fd_attach_probe(struct fd_bpf *obj)
{
obj->links.netdata_sys_open_kprobe = bpf_program__attach_kprobe(obj->progs.netdata_sys_open_kprobe, false,
fd_targets[NETDATA_FD_SYSCALL_OPEN].name);
- int ret = libbpf_get_error(obj->links.netdata_sys_open_kprobe);
+ long ret = libbpf_get_error(obj->links.netdata_sys_open_kprobe);
if (ret)
return -1;
@@ -187,13 +191,6 @@ static int ebpf_fd_attach_probe(struct fd_bpf *obj)
if (ret)
return -1;
- obj->links.netdata_release_task_fd_kprobe = bpf_program__attach_kprobe(obj->progs.netdata_release_task_fd_kprobe,
- false,
- EBPF_COMMON_FNCT_CLEAN_UP);
- ret = libbpf_get_error(obj->links.netdata_release_task_fd_kprobe);
- if (ret)
- return -1;
-
if (!strcmp(fd_targets[NETDATA_FD_SYSCALL_CLOSE].name, close_targets[NETDATA_FD_CLOSE_FD])) {
obj->links.netdata_close_fd_kretprobe = bpf_program__attach_kprobe(obj->progs.netdata_close_fd_kretprobe, true,
fd_targets[NETDATA_FD_SYSCALL_CLOSE].name);
@@ -302,19 +299,6 @@ static void ebpf_fd_adjust_map(struct fd_bpf *obj, ebpf_module_t *em)
}
/**
- * Disable Release Task
- *
- * Disable release task when apps is not enabled.
- *
- * @param obj is the main structure for bpf objects.
- */
-static void ebpf_fd_disable_release_task(struct fd_bpf *obj)
-{
- bpf_program__set_autoload(obj->progs.netdata_release_task_fd_kprobe, false);
- bpf_program__set_autoload(obj->progs.netdata_release_task_fd_fentry, false);
-}
-
-/**
* Load and attach
*
* Load and attach the eBPF code in kernel.
@@ -339,8 +323,6 @@ static inline int ebpf_fd_load_and_attach(struct fd_bpf *obj, ebpf_module_t *em)
ebpf_disable_specific_trampoline(obj);
ebpf_set_trampoline_target(obj);
- // TODO: Remove this in next PR, because this specific trampoline has an error.
- bpf_program__set_autoload(obj->progs.netdata_release_task_fd_fentry, false);
} else {
ebpf_disable_trampoline(obj);
ebpf_disable_specific_probes(obj);
@@ -348,9 +330,6 @@ static inline int ebpf_fd_load_and_attach(struct fd_bpf *obj, ebpf_module_t *em)
ebpf_fd_adjust_map(obj, em);
- if (!em->apps_charts && !em->cgroup_charts)
- ebpf_fd_disable_release_task(obj);
-
int ret = fd_bpf__load(obj);
if (ret) {
return ret;
@@ -382,14 +361,14 @@ static void ebpf_obsolete_specific_fd_charts(char *type, ebpf_module_t *em);
*
* @param em a pointer to `struct ebpf_module`
*/
-static void ebpf_obsolete_fd_services(ebpf_module_t *em)
+static void ebpf_obsolete_fd_services(ebpf_module_t *em, char *id)
{
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_FILE_OPEN,
- "",
"Number of open files",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_FILE_CGROUP_GROUP,
+ NETDATA_APPS_FILE_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NETDATA_CGROUP_FD_OPEN_CONTEXT,
20270,
@@ -397,11 +376,11 @@ static void ebpf_obsolete_fd_services(ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_FILE_OPEN_ERROR,
- "",
"Fails to open files",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_FILE_CGROUP_GROUP,
+ NETDATA_APPS_FILE_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NETDATA_CGROUP_FD_OPEN_ERR_CONTEXT,
20271,
@@ -409,11 +388,11 @@ static void ebpf_obsolete_fd_services(ebpf_module_t *em)
}
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_FILE_CLOSED,
- "",
"Files closed",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_FILE_CGROUP_GROUP,
+ NETDATA_APPS_FILE_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NETDATA_CGROUP_FD_CLOSE_CONTEXT,
20272,
@@ -421,11 +400,11 @@ static void ebpf_obsolete_fd_services(ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_FILE_CLOSE_ERROR,
- "",
"Fails to close files",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_FILE_CGROUP_GROUP,
+ NETDATA_APPS_FILE_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NETDATA_CGROUP_FD_CLOSE_ERR_CONTEXT,
20273,
@@ -443,12 +422,13 @@ static void ebpf_obsolete_fd_services(ebpf_module_t *em)
static inline void ebpf_obsolete_fd_cgroup_charts(ebpf_module_t *em) {
pthread_mutex_lock(&mutex_cgroup_shm);
- ebpf_obsolete_fd_services(em);
-
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (ect->systemd)
+ if (ect->systemd) {
+ ebpf_obsolete_fd_services(em, ect->name);
+
continue;
+ }
ebpf_obsolete_specific_fd_charts(ect->name, em);
}
@@ -466,6 +446,7 @@ void ebpf_obsolete_fd_apps_charts(struct ebpf_module *em)
{
struct ebpf_target *w;
int update_every = em->update_every;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = apps_groups_root_target; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_FD_IDX))))
continue;
@@ -519,6 +500,7 @@ void ebpf_obsolete_fd_apps_charts(struct ebpf_module *em)
}
w->charts_created &= ~(1<<EBPF_MODULE_FD_IDX);
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -566,6 +548,9 @@ static void ebpf_fd_exit(void *ptr)
{
ebpf_module_t *em = (ebpf_module_t *)ptr;
+ if (ebpf_read_fd.thread)
+ netdata_thread_cancel(*ebpf_read_fd.thread);
+
if (em->enabled == NETDATA_THREAD_EBPF_FUNCTION_RUNNING) {
pthread_mutex_lock(&lock);
if (em->cgroup_charts) {
@@ -579,12 +564,6 @@ static void ebpf_fd_exit(void *ptr)
ebpf_obsolete_fd_global(em);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_fd_pid)
- ebpf_statistic_obsolete_aral_chart(em, fd_disable_priority);
-#endif
-
-
fflush(stdout);
pthread_mutex_unlock(&lock);
}
@@ -684,73 +663,152 @@ static void fd_apps_accumulator(netdata_fd_stat_t *out, int maps_per_core)
}
/**
- * Fill PID
- *
- * Fill PID structures
- *
- * @param current_pid pid that we are collecting data
- * @param out values read from hash tables;
- */
-static void fd_fill_pid(uint32_t current_pid, netdata_fd_stat_t *publish)
-{
- netdata_fd_stat_t *curr = fd_pid[current_pid];
- if (!curr) {
- curr = ebpf_fd_stat_get();
- fd_pid[current_pid] = curr;
- }
-
- memcpy(curr, &publish[0], sizeof(netdata_fd_stat_t));
-}
-
-/**
* Read APPS table
*
* Read the apps table and store data inside the structure.
*
* @param maps_per_core do I need to read all cores?
*/
-static void read_fd_apps_table(int maps_per_core)
+static void ebpf_read_fd_apps_table(int maps_per_core, int max_period)
{
netdata_fd_stat_t *fv = fd_vector;
- uint32_t key;
- struct ebpf_pid_stat *pids = ebpf_root_of_pids;
int fd = fd_maps[NETDATA_FD_PID_STATS].map_fd;
size_t length = sizeof(netdata_fd_stat_t);
if (maps_per_core)
length *= ebpf_nprocs;
- while (pids) {
- key = pids->pid;
-
+ uint32_t key = 0, next_key = 0;
+ while (bpf_map_get_next_key(fd, &key, &next_key) == 0) {
if (bpf_map_lookup_elem(fd, &key, fv)) {
- pids = pids->next;
- continue;
+ goto end_fd_loop;
}
fd_apps_accumulator(fv, maps_per_core);
- fd_fill_pid(key, fv);
+ ebpf_pid_stat_t *pid_stat = ebpf_get_pid_entry(key, fv->tgid);
+ if (pid_stat) {
+ netdata_fd_stat_t *publish_fd = &pid_stat->fd;
+ if (!publish_fd->ct || publish_fd->ct != fv->ct) {
+ memcpy(publish_fd, &fv[0], sizeof(netdata_fd_stat_t));
+ pid_stat->not_updated = 0;
+ } else if (++pid_stat->not_updated >= max_period) {
+ bpf_map_delete_elem(fd, &key);
+ pid_stat->not_updated = 0;
+ }
+ }
+end_fd_loop:
// We are cleaning to avoid passing data read from one process to other.
memset(fv, 0, length);
+ key = next_key;
+ }
+}
- pids = pids->next;
+/**
+ * Sum PIDs
+ *
+ * Sum values for all targets.
+ *
+ * @param fd the output
+ * @param root list of pids
+ */
+static void ebpf_fd_sum_pids(netdata_fd_stat_t *fd, struct ebpf_pid_on_target *root)
+{
+ memset(fd, 0, sizeof(netdata_fd_stat_t));
+
+ while (root) {
+ int32_t pid = root->pid;
+ ebpf_pid_stat_t *pid_stat = ebpf_get_pid_entry(pid, 0);
+ if (pid_stat) {
+ netdata_fd_stat_t *w = &pid_stat->fd;
+ fd->open_call += w->open_call;
+ fd->close_call += w->close_call;
+ fd->open_err += w->open_err;
+ fd->close_err += w->close_err;
+ }
+
+ root = root->next;
+ }
+}
+
+/**
+ * Resume apps data
+ */
+void ebpf_fd_resume_apps_data()
+{
+ struct ebpf_target *w;
+
+ for (w = apps_groups_root_target; w; w = w->next) {
+ if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_FD_IDX))))
+ continue;
+
+ ebpf_fd_sum_pids(&w->fd, w->root_pid);
}
}
/**
+ * DCstat thread
+ *
+ * Thread used to generate dcstat charts.
+ *
+ * @param ptr a pointer to `struct ebpf_module`
+ *
+ * @return It always return NULL
+ */
+void *ebpf_read_fd_thread(void *ptr)
+{
+ heartbeat_t hb;
+ heartbeat_init(&hb);
+
+ ebpf_module_t *em = (ebpf_module_t *)ptr;
+
+ int maps_per_core = em->maps_per_core;
+ int update_every = em->update_every;
+
+ int counter = update_every - 1;
+
+ uint32_t lifetime = em->lifetime;
+ uint32_t running_time = 0;
+ usec_t period = update_every * USEC_PER_SEC;
+ int max_period = update_every * EBPF_CLEANUP_FACTOR;
+ while (!ebpf_plugin_exit && running_time < lifetime) {
+ (void)heartbeat_next(&hb, period);
+ if (ebpf_plugin_exit || ++counter != update_every)
+ continue;
+
+ netdata_thread_disable_cancelability();
+
+ pthread_mutex_lock(&collect_data_mutex);
+ ebpf_read_fd_apps_table(maps_per_core, max_period);
+ ebpf_fd_resume_apps_data();
+ pthread_mutex_unlock(&collect_data_mutex);
+
+ counter = 0;
+
+ pthread_mutex_lock(&ebpf_exit_cleanup);
+ if (running_time && !em->running_time)
+ running_time = update_every;
+ else
+ running_time += update_every;
+
+ em->running_time = running_time;
+ pthread_mutex_unlock(&ebpf_exit_cleanup);
+ netdata_thread_enable_cancelability();
+ }
+
+ return NULL;
+}
+
+/**
* Update cgroup
*
* Update cgroup data collected per PID.
*
* @param maps_per_core do I need to read all cores?
*/
-static void ebpf_update_fd_cgroup(int maps_per_core)
+static void ebpf_update_fd_cgroup()
{
ebpf_cgroup_target_t *ect ;
- netdata_fd_stat_t *fv = fd_vector;
- int fd = fd_maps[NETDATA_FD_PID_STATS].map_fd;
- size_t length = sizeof(netdata_fd_stat_t) * ebpf_nprocs;
pthread_mutex_lock(&mutex_cgroup_shm);
for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
@@ -758,17 +816,11 @@ static void ebpf_update_fd_cgroup(int maps_per_core)
for (pids = ect->pids; pids; pids = pids->next) {
int pid = pids->pid;
netdata_fd_stat_t *out = &pids->fd;
- if (likely(fd_pid) && fd_pid[pid]) {
- netdata_fd_stat_t *in = fd_pid[pid];
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ netdata_fd_stat_t *in = &local_pid->fd;
memcpy(out, in, sizeof(netdata_fd_stat_t));
- } else {
- memset(fv, 0, length);
- if (!bpf_map_lookup_elem(fd, &pid, fv)) {
- fd_apps_accumulator(fv, maps_per_core);
-
- memcpy(out, fv, sizeof(netdata_fd_stat_t));
- }
}
}
}
@@ -776,41 +828,6 @@ static void ebpf_update_fd_cgroup(int maps_per_core)
}
/**
- * Sum PIDs
- *
- * Sum values for all targets.
- *
- * @param fd the output
- * @param root list of pids
- */
-static void ebpf_fd_sum_pids(netdata_fd_stat_t *fd, struct ebpf_pid_on_target *root)
-{
- uint32_t open_call = 0;
- uint32_t close_call = 0;
- uint32_t open_err = 0;
- uint32_t close_err = 0;
-
- while (root) {
- int32_t pid = root->pid;
- netdata_fd_stat_t *w = fd_pid[pid];
- if (w) {
- open_call += w->open_call;
- close_call += w->close_call;
- open_err += w->open_err;
- close_err += w->close_err;
- }
-
- root = root->next;
- }
-
- // These conditions were added, because we are using incremental algorithm
- fd->open_call = (open_call >= fd->open_call) ? open_call : fd->open_call;
- fd->close_call = (close_call >= fd->close_call) ? close_call : fd->close_call;
- fd->open_err = (open_err >= fd->open_err) ? open_err : fd->open_err;
- fd->close_err = (close_err >= fd->close_err) ? close_err : fd->close_err;
-}
-
-/**
* Send data to Netdata calling auxiliary functions.
*
* @param em the structure with thread information
@@ -819,12 +836,11 @@ static void ebpf_fd_sum_pids(netdata_fd_stat_t *fd, struct ebpf_pid_on_target *r
void ebpf_fd_send_apps_data(ebpf_module_t *em, struct ebpf_target *root)
{
struct ebpf_target *w;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = root; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_FD_IDX))))
continue;
- ebpf_fd_sum_pids(&w->fd, w->root_pid);
-
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_file_open");
write_chart_dimension("calls", w->fd.open_call);
ebpf_write_end_chart();
@@ -845,6 +861,7 @@ void ebpf_fd_send_apps_data(ebpf_module_t *em, struct ebpf_target *root)
ebpf_write_end_chart();
}
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -888,7 +905,7 @@ static void ebpf_fd_sum_cgroup_pids(netdata_fd_stat_t *fd, struct pid_on_target2
static void ebpf_create_specific_fd_charts(char *type, ebpf_module_t *em)
{
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_FILE_OPEN, "Number of open files",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_CGROUP_GROUP,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_GROUP,
NETDATA_CGROUP_FD_OPEN_CONTEXT, NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5400,
ebpf_create_global_dimension,
@@ -897,7 +914,7 @@ static void ebpf_create_specific_fd_charts(char *type, ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_FILE_OPEN_ERROR, "Fails to open files",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_CGROUP_GROUP,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_GROUP,
NETDATA_CGROUP_FD_OPEN_ERR_CONTEXT, NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5401,
ebpf_create_global_dimension,
@@ -907,7 +924,7 @@ static void ebpf_create_specific_fd_charts(char *type, ebpf_module_t *em)
}
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_FILE_CLOSED, "Files closed",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_CGROUP_GROUP,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_GROUP,
NETDATA_CGROUP_FD_CLOSE_CONTEXT, NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5402,
ebpf_create_global_dimension,
@@ -916,7 +933,7 @@ static void ebpf_create_specific_fd_charts(char *type, ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_FILE_CLOSE_ERROR, "Fails to close files",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_CGROUP_GROUP,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_GROUP,
NETDATA_CGROUP_FD_CLOSE_ERR_CONTEXT, NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5403,
ebpf_create_global_dimension,
@@ -1001,32 +1018,84 @@ static void ebpf_send_specific_fd_data(char *type, netdata_fd_stat_t *values, eb
**/
static void ebpf_create_systemd_fd_charts(ebpf_module_t *em)
{
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_FILE_OPEN, "Number of open files",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20061,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_FD_OPEN_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_FD, em->update_every);
+ static ebpf_systemd_args_t data_open = {
+ .title = "Number of open files",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_FILE_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20061,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_FD_OPEN_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_FD,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_FILE_OPEN,
+ .dimension = "calls"
+ };
- if (em->mode < MODE_ENTRY) {
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_FILE_OPEN_ERROR, "Fails to open files",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20062,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_FD_OPEN_ERR_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_FD, em->update_every);
- }
+ static ebpf_systemd_args_t data_open_error = {
+ .title = "Fails to open files",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_FILE_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20062,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_FD_OPEN_ERR_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_FD,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_FILE_OPEN_ERROR,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_FILE_CLOSED, "Files closed",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20063,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_FD_CLOSE_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_FD, em->update_every);
+ static ebpf_systemd_args_t data_close = {
+ .title = "Files closed",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_FILE_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20063,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_FD_CLOSE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_FD,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_FILE_CLOSED,
+ .dimension = "calls"
+ };
- if (em->mode < MODE_ENTRY) {
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_FILE_CLOSE_ERROR, "Fails to close files",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_FILE_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20064,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_FD_CLOSE_ERR_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_FD, em->update_every);
+ static ebpf_systemd_args_t data_close_error = {
+ .title = "Fails to close files",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_FILE_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20064,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_FD_OPEN_ERR_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_FD,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_FILE_CLOSE_ERROR,
+ .dimension = "calls"
+ };
+
+ if (!data_open.update_every)
+ data_open.update_every = data_open_error.update_every =
+ data_close.update_every = data_close_error.update_every = em->update_every;
+
+ ebpf_cgroup_target_t *w;
+ netdata_run_mode_t mode = em->mode;
+ for (w = ebpf_cgroup_pids; w; w = w->next) {
+ if (unlikely(!w->systemd || w->flags & NETDATA_EBPF_SERVICES_HAS_FD_CHART))
+ continue;
+
+ data_open.id = data_open_error.id = data_close.id = data_close_error.id = w->name;
+ ebpf_create_charts_on_systemd(&data_open);
+ if (mode < MODE_ENTRY) {
+ ebpf_create_charts_on_systemd(&data_open_error);
+ }
+
+ ebpf_create_charts_on_systemd(&data_close);
+ if (mode < MODE_ENTRY) {
+ ebpf_create_charts_on_systemd(&data_close_error);
+ }
+
+ w->flags |= NETDATA_EBPF_SERVICES_HAS_FD_CHART;
}
}
@@ -1040,40 +1109,30 @@ static void ebpf_create_systemd_fd_charts(ebpf_module_t *em)
static void ebpf_send_systemd_fd_charts(ebpf_module_t *em)
{
ebpf_cgroup_target_t *ect;
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_FILE_OPEN, "");
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_fd.open_call);
+ if (unlikely(!(ect->flags & NETDATA_EBPF_SERVICES_HAS_FD_CHART)) ) {
+ continue;
}
- }
- ebpf_write_end_chart();
- if (em->mode < MODE_ENTRY) {
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_FILE_OPEN_ERROR, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_fd.open_err);
- }
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_FILE_OPEN);
+ write_chart_dimension("calls", ect->publish_systemd_fd.open_call);
ebpf_write_end_chart();
- }
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_FILE_CLOSED, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_fd.close_call);
+ if (em->mode < MODE_ENTRY) {
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_FILE_OPEN_ERROR);
+ write_chart_dimension("calls", ect->publish_systemd_fd.open_err);
+ ebpf_write_end_chart();
}
- }
- ebpf_write_end_chart();
- if (em->mode < MODE_ENTRY) {
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_FILE_CLOSE_ERROR, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_fd.close_err);
- }
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_FILE_CLOSED);
+ write_chart_dimension("calls", ect->publish_systemd_fd.close_call);
ebpf_write_end_chart();
+
+ if (em->mode < MODE_ENTRY) {
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_FILE_CLOSE_ERROR);
+ write_chart_dimension("calls", ect->publish_systemd_fd.close_err);
+ ebpf_write_end_chart();
+ }
}
}
@@ -1084,17 +1143,13 @@ static void ebpf_send_systemd_fd_charts(ebpf_module_t *em)
*/
static void ebpf_fd_send_cgroup_data(ebpf_module_t *em)
{
- if (!ebpf_cgroup_pids)
- return;
-
pthread_mutex_lock(&mutex_cgroup_shm);
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
ebpf_fd_sum_cgroup_pids(&ect->publish_systemd_fd, ect->pids);
}
- int has_systemd = shm_ebpf_cgroup.header->systemd_enabled;
- if (has_systemd) {
+ if (shm_ebpf_cgroup.header->systemd_enabled) {
if (send_cgroup_chart) {
ebpf_create_systemd_fd_charts(em);
}
@@ -1148,30 +1203,21 @@ static void fd_collector(ebpf_module_t *em)
counter = 0;
netdata_apps_integration_flags_t apps = em->apps_charts;
ebpf_fd_read_global_tables(stats, maps_per_core);
- pthread_mutex_lock(&collect_data_mutex);
- if (apps)
- read_fd_apps_table(maps_per_core);
if (cgroups)
- ebpf_update_fd_cgroup(maps_per_core);
+ ebpf_update_fd_cgroup();
pthread_mutex_lock(&lock);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_fd_pid)
- ebpf_send_data_aral_chart(ebpf_aral_fd_pid, em);
-#endif
-
ebpf_fd_send_data(em);
if (apps & NETDATA_EBPF_APPS_FLAG_CHART_CREATED)
ebpf_fd_send_apps_data(em, apps_groups_root_target);
- if (cgroups)
+ if (cgroups && shm_ebpf_cgroup.header && ebpf_cgroup_pids)
ebpf_fd_send_cgroup_data(em);
pthread_mutex_unlock(&lock);
- pthread_mutex_unlock(&collect_data_mutex);
pthread_mutex_lock(&ebpf_exit_cleanup);
if (running_time && !em->running_time)
@@ -1327,17 +1373,10 @@ static void ebpf_create_fd_global_charts(ebpf_module_t *em)
*
* We are not testing the return, because callocz does this and shutdown the software
* case it was not possible to allocate.
- *
- * @param apps is apps enabled?
*/
-static void ebpf_fd_allocate_global_vectors(int apps)
+static inline void ebpf_fd_allocate_global_vectors()
{
- if (apps) {
- ebpf_fd_aral_init();
- fd_pid = callocz((size_t)pid_max, sizeof(netdata_fd_stat_t *));
- fd_vector = callocz((size_t)ebpf_nprocs, sizeof(netdata_fd_stat_t));
- }
-
+ fd_vector = callocz((size_t)ebpf_nprocs, sizeof(netdata_fd_stat_t));
fd_values = callocz((size_t)ebpf_nprocs, sizeof(netdata_idx_t));
}
@@ -1401,7 +1440,7 @@ void *ebpf_fd_thread(void *ptr)
goto endfd;
}
- ebpf_fd_allocate_global_vectors(em->apps_charts);
+ ebpf_fd_allocate_global_vectors();
int algorithms[NETDATA_FD_SYSCALL_END] = {
NETDATA_EBPF_INCREMENTAL_IDX, NETDATA_EBPF_INCREMENTAL_IDX
@@ -1414,13 +1453,16 @@ void *ebpf_fd_thread(void *ptr)
ebpf_create_fd_global_charts(em);
ebpf_update_stats(&plugin_statistics, em);
ebpf_update_kernel_memory_with_vector(&plugin_statistics, em->maps, EBPF_ACTION_STAT_ADD);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_fd_pid)
- fd_disable_priority = ebpf_statistic_create_aral_chart(NETDATA_EBPF_FD_ARAL_NAME, em);
-#endif
pthread_mutex_unlock(&lock);
+ ebpf_read_fd.thread = mallocz(sizeof(netdata_thread_t));
+ netdata_thread_create(ebpf_read_fd.thread,
+ ebpf_read_fd.name,
+ NETDATA_THREAD_OPTION_DEFAULT,
+ ebpf_read_fd_thread,
+ em);
+
fd_collector(em);
endfd:
diff --git a/collectors/ebpf.plugin/ebpf_fd.h b/src/collectors/ebpf.plugin/ebpf_fd.h
index 00986673e..65658d9b3 100644
--- a/collectors/ebpf.plugin/ebpf_fd.h
+++ b/src/collectors/ebpf.plugin/ebpf_fd.h
@@ -29,15 +29,21 @@
#define NETDATA_CGROUP_FD_CLOSE_CONTEXT "cgroup.fd_close"
#define NETDATA_CGROUP_FD_CLOSE_ERR_CONTEXT "cgroup.fd_close_error"
-#define NETDATA_SYSTEMD_FD_OPEN_CONTEXT "services.fd_open"
-#define NETDATA_SYSTEMD_FD_OPEN_ERR_CONTEXT "services.fd_open_error"
-#define NETDATA_SYSTEMD_FD_CLOSE_CONTEXT "services.fd_close"
-#define NETDATA_SYSTEMD_FD_CLOSE_ERR_CONTEXT "services.fd_close_error"
+#define NETDATA_SYSTEMD_FD_OPEN_CONTEXT "systemd.services.fd_open"
+#define NETDATA_SYSTEMD_FD_OPEN_ERR_CONTEXT "systemd.services.fd_open_error"
+#define NETDATA_SYSTEMD_FD_CLOSE_CONTEXT "systemd.services.fd_close"
+#define NETDATA_SYSTEMD_FD_CLOSE_ERR_CONTEXT "systemd.services.fd_close_error"
// ARAL name
#define NETDATA_EBPF_FD_ARAL_NAME "ebpf_fd"
typedef struct netdata_fd_stat {
+ uint64_t ct;
+ uint32_t tgid;
+ uint32_t uid;
+ uint32_t gid;
+ char name[TASK_COMM_LEN];
+
uint32_t open_call; // Open syscalls (open and openat)
uint32_t close_call; // Close syscall (close)
diff --git a/collectors/ebpf.plugin/ebpf_filesystem.c b/src/collectors/ebpf.plugin/ebpf_filesystem.c
index b78e65532..b5c898232 100644
--- a/collectors/ebpf.plugin/ebpf_filesystem.c
+++ b/src/collectors/ebpf.plugin/ebpf_filesystem.c
@@ -321,7 +321,7 @@ static inline int ebpf_fs_load_and_attach(ebpf_local_maps_t *map, struct filesys
ret = ebpf_fs_attach_kprobe(obj, functions);
if (!ret)
- map->map_fd = bpf_map__fd(obj->maps.tbl_fs);;
+ map->map_fd = bpf_map__fd(obj->maps.tbl_fs);
return ret;
}
diff --git a/collectors/ebpf.plugin/ebpf_filesystem.h b/src/collectors/ebpf.plugin/ebpf_filesystem.h
index f58d7fbe4..cd54be57e 100644
--- a/collectors/ebpf.plugin/ebpf_filesystem.h
+++ b/src/collectors/ebpf.plugin/ebpf_filesystem.h
@@ -8,9 +8,6 @@
#define NETDATA_EBPF_FS_MODULE_DESC "Monitor filesystem latency for: btrfs, ext4, nfs, xfs and zfs."
#include "ebpf.h"
-#ifdef LIBBPF_MAJOR_VERSION
-#include "includes/filesystem.skel.h"
-#endif
#define NETDATA_FS_MAX_DIST_NAME 64UL
diff --git a/collectors/ebpf.plugin/ebpf_hardirq.c b/src/collectors/ebpf.plugin/ebpf_hardirq.c
index 465ee6434..465ee6434 100644
--- a/collectors/ebpf.plugin/ebpf_hardirq.c
+++ b/src/collectors/ebpf.plugin/ebpf_hardirq.c
diff --git a/collectors/ebpf.plugin/ebpf_hardirq.h b/src/collectors/ebpf.plugin/ebpf_hardirq.h
index 35b03b761..35b03b761 100644
--- a/collectors/ebpf.plugin/ebpf_hardirq.h
+++ b/src/collectors/ebpf.plugin/ebpf_hardirq.h
diff --git a/collectors/ebpf.plugin/ebpf_mdflush.c b/src/collectors/ebpf.plugin/ebpf_mdflush.c
index fe33ff6a4..fe33ff6a4 100644
--- a/collectors/ebpf.plugin/ebpf_mdflush.c
+++ b/src/collectors/ebpf.plugin/ebpf_mdflush.c
diff --git a/collectors/ebpf.plugin/ebpf_mdflush.h b/src/collectors/ebpf.plugin/ebpf_mdflush.h
index 629550746..629550746 100644
--- a/collectors/ebpf.plugin/ebpf_mdflush.h
+++ b/src/collectors/ebpf.plugin/ebpf_mdflush.h
diff --git a/collectors/ebpf.plugin/ebpf_mount.c b/src/collectors/ebpf.plugin/ebpf_mount.c
index 05c76540a..05c76540a 100644
--- a/collectors/ebpf.plugin/ebpf_mount.c
+++ b/src/collectors/ebpf.plugin/ebpf_mount.c
diff --git a/collectors/ebpf.plugin/ebpf_mount.h b/src/collectors/ebpf.plugin/ebpf_mount.h
index 768914b02..768914b02 100644
--- a/collectors/ebpf.plugin/ebpf_mount.h
+++ b/src/collectors/ebpf.plugin/ebpf_mount.h
diff --git a/collectors/ebpf.plugin/ebpf_oomkill.c b/src/collectors/ebpf.plugin/ebpf_oomkill.c
index 2c34650c3..e7604a2db 100644
--- a/collectors/ebpf.plugin/ebpf_oomkill.c
+++ b/src/collectors/ebpf.plugin/ebpf_oomkill.c
@@ -53,11 +53,11 @@ static void ebpf_create_specific_oomkill_charts(char *type, int update_every);
*
* @param em a pointer to `struct ebpf_module`
*/
-static void ebpf_obsolete_oomkill_services(ebpf_module_t *em)
+static void ebpf_obsolete_oomkill_services(ebpf_module_t *em, char *id)
{
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_OOMKILL_CHART,
- "",
"OOM kills. This chart is provided by eBPF plugin.",
EBPF_COMMON_DIMENSION_KILLS,
NETDATA_EBPF_MEMORY_GROUP,
@@ -78,12 +78,13 @@ static inline void ebpf_obsolete_oomkill_cgroup_charts(ebpf_module_t *em)
{
pthread_mutex_lock(&mutex_cgroup_shm);
- ebpf_obsolete_oomkill_services(em);
-
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (ect->systemd)
+ if (ect->systemd) {
+ ebpf_obsolete_oomkill_services(em, ect->name);
+
continue;
+ }
ebpf_create_specific_oomkill_charts(ect->name, em->update_every);
}
@@ -101,6 +102,7 @@ static void ebpf_obsolete_oomkill_apps(ebpf_module_t *em)
{
struct ebpf_target *w;
int update_every = em->update_every;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = apps_groups_root_target; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_OOMKILL_IDX))))
continue;
@@ -118,6 +120,7 @@ static void ebpf_obsolete_oomkill_apps(ebpf_module_t *em)
w->charts_created &= ~(1<<EBPF_MODULE_OOMKILL_IDX);
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -231,11 +234,33 @@ static void ebpf_create_specific_oomkill_charts(char *type, int update_every)
**/
static void ebpf_create_systemd_oomkill_charts(int update_every)
{
- ebpf_create_charts_on_systemd(NETDATA_OOMKILL_CHART, "OOM kills. This chart is provided by eBPF plugin.",
- EBPF_COMMON_DIMENSION_KILLS, NETDATA_EBPF_MEMORY_GROUP,
- NETDATA_EBPF_CHART_TYPE_LINE, 20191,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NULL,
- NETDATA_EBPF_MODULE_NAME_OOMKILL, update_every);
+ static ebpf_systemd_args_t data_oom = {
+ .title = "OOM kills. This chart is provided by eBPF plugin.",
+ .units = EBPF_COMMON_DIMENSION_KILLS,
+ .family = NETDATA_EBPF_MEMORY_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20191,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_EBPF_MODULE_NAME_OOMKILL,
+ .module = NETDATA_EBPF_MODULE_NAME_SWAP,
+ .update_every = 0,
+ .suffix = NETDATA_OOMKILL_CHART,
+ .dimension = "oom"
+ };
+
+ if (!data_oom.update_every)
+ data_oom.update_every = update_every;
+
+ ebpf_cgroup_target_t *w;
+ for (w = ebpf_cgroup_pids; w ; w = w->next) {
+ if (unlikely(!w->systemd || w->flags & NETDATA_EBPF_SERVICES_HAS_OOMKILL_CHART))
+ continue;
+
+ data_oom.id = w->name;
+ ebpf_create_charts_on_systemd(&data_oom);
+
+ w->flags |= NETDATA_EBPF_SERVICES_HAS_OOMKILL_CHART;
+ }
}
/**
@@ -246,14 +271,15 @@ static void ebpf_create_systemd_oomkill_charts(int update_every)
static void ebpf_send_systemd_oomkill_charts()
{
ebpf_cgroup_target_t *ect;
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_OOMKILL_CHART, "");
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long) ect->oomkill);
- ect->oomkill = 0;
+ if (unlikely(!(ect->flags & NETDATA_EBPF_SERVICES_HAS_OOMKILL_CHART)) ) {
+ continue;
}
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_OOMKILL_CHART);
+ write_chart_dimension("oom", (long long) ect->oomkill);
+ ect->oomkill = 0;
+ ebpf_write_end_chart();
}
- ebpf_write_end_chart();
}
/*
@@ -294,14 +320,10 @@ static void ebpf_obsolete_specific_oomkill_charts(char *type, int update_every)
*/
void ebpf_oomkill_send_cgroup_data(int update_every)
{
- if (!ebpf_cgroup_pids)
- return;
-
pthread_mutex_lock(&mutex_cgroup_shm);
ebpf_cgroup_target_t *ect;
- int has_systemd = shm_ebpf_cgroup.header->systemd_enabled;
- if (has_systemd) {
+ if (shm_ebpf_cgroup.header->systemd_enabled) {
if (send_cgroup_chart) {
ebpf_create_systemd_oomkill_charts(update_every);
}
@@ -443,26 +465,23 @@ static void oomkill_collector(ebpf_module_t *em)
counter = 0;
uint32_t count = oomkill_read_data(keys);
- if (!count) {
- running_time = ebpf_update_oomkill_period(running_time, em);
- }
stats[NETDATA_CONTROLLER_PID_TABLE_ADD] += (uint64_t) count;
stats[NETDATA_CONTROLLER_PID_TABLE_DEL] += (uint64_t) count;
- pthread_mutex_lock(&collect_data_mutex);
pthread_mutex_lock(&lock);
- if (cgroups && count) {
+ if (cgroups && shm_ebpf_cgroup.header && ebpf_cgroup_pids) {
ebpf_update_oomkill_cgroup(keys, count);
// write everything from the ebpf map.
ebpf_oomkill_send_cgroup_data(update_every);
}
if (em->apps_charts & NETDATA_EBPF_APPS_FLAG_CHART_CREATED) {
+ pthread_mutex_lock(&collect_data_mutex);
oomkill_write_data(keys, count);
+ pthread_mutex_unlock(&collect_data_mutex);
}
pthread_mutex_unlock(&lock);
- pthread_mutex_unlock(&collect_data_mutex);
running_time = ebpf_update_oomkill_period(running_time, em);
}
diff --git a/collectors/ebpf.plugin/ebpf_oomkill.h b/src/collectors/ebpf.plugin/ebpf_oomkill.h
index 4a5fa62aa..4a5fa62aa 100644
--- a/collectors/ebpf.plugin/ebpf_oomkill.h
+++ b/src/collectors/ebpf.plugin/ebpf_oomkill.h
diff --git a/collectors/ebpf.plugin/ebpf_process.c b/src/collectors/ebpf.plugin/ebpf_process.c
index e3e2b884e..306b59639 100644
--- a/collectors/ebpf.plugin/ebpf_process.c
+++ b/src/collectors/ebpf.plugin/ebpf_process.c
@@ -1,7 +1,5 @@
// SPDX-License-Identifier: GPL-3.0-or-later
-#include <sys/resource.h>
-
#include "ebpf.h"
#include "ebpf_process.h"
@@ -65,10 +63,6 @@ struct config process_config = { .first_section = NULL,
.index = { .avl_tree = { .root = NULL, .compar = appconfig_section_compare },
.rwlock = AVL_LOCK_INITIALIZER } };
-#ifdef NETDATA_DEV_MODE
-int process_disable_priority;
-#endif
-
/*****************************************************************
*
* PROCESS DATA AND SEND TO NETDATA
@@ -147,53 +141,6 @@ static void ebpf_process_send_data(ebpf_module_t *em)
}
/**
- * Sum values for pid
- *
- * @param root the structure with all available PIDs
- * @param offset the address that we are reading
- *
- * @return it returns the sum of all PIDs
- */
-long long ebpf_process_sum_values_for_pids(struct ebpf_pid_on_target *root, size_t offset)
-{
- long long ret = 0;
- while (root) {
- int32_t pid = root->pid;
- ebpf_process_stat_t *w = global_process_stats[pid];
- if (w) {
- uint32_t *value = (uint32_t *)((char *)w + offset);
- ret += *value;
- }
-
- root = root->next;
- }
-
- return ret;
-}
-
-/**
- * Remove process pid
- *
- * Remove from PID task table when task_release was called.
- */
-void ebpf_process_remove_pids()
-{
- struct ebpf_pid_stat *pids = ebpf_root_of_pids;
- int pid_fd = process_maps[NETDATA_PROCESS_PID_TABLE].map_fd;
- while (pids) {
- uint32_t pid = pids->pid;
- ebpf_process_stat_t *w = global_process_stats[pid];
- if (w) {
- ebpf_process_stat_release(w);
- global_process_stats[pid] = NULL;
- bpf_map_delete_elem(pid_fd, &pid);
- }
-
- pids = pids->next;
- }
-}
-
-/**
* Send data to Netdata calling auxiliary functions.
*
* @param root the target list.
@@ -201,46 +148,33 @@ void ebpf_process_remove_pids()
void ebpf_process_send_apps_data(struct ebpf_target *root, ebpf_module_t *em)
{
struct ebpf_target *w;
- // This algorithm is improved in https://github.com/netdata/netdata/pull/16030
- collected_number values[5];
for (w = root; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_PROCESS_IDX))))
continue;
- values[0] = ebpf_process_sum_values_for_pids(w->root_pid, offsetof(ebpf_process_stat_t, create_process));
- values[1] = ebpf_process_sum_values_for_pids(w->root_pid, offsetof(ebpf_process_stat_t, create_thread));
- values[2] = ebpf_process_sum_values_for_pids(w->root_pid, offsetof(ebpf_process_stat_t,
- exit_call));
- values[3] = ebpf_process_sum_values_for_pids(w->root_pid, offsetof(ebpf_process_stat_t,
- release_call));
- values[4] = ebpf_process_sum_values_for_pids(w->root_pid, offsetof(ebpf_process_stat_t,
- task_err));
-
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_process_start");
- write_chart_dimension("calls", values[0]);
+ write_chart_dimension("calls", w->process.create_process);
ebpf_write_end_chart();
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_thread_start");
- write_chart_dimension("calls", values[1]);
+ write_chart_dimension("calls", w->process.create_thread);
ebpf_write_end_chart();
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_task_exit");
- write_chart_dimension("calls", values[2]);
+ write_chart_dimension("calls", w->process.exit_call);
ebpf_write_end_chart();
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_task_released");
- write_chart_dimension("calls", values[3]);
+ write_chart_dimension("calls", w->process.release_call);
ebpf_write_end_chart();
if (em->mode < MODE_ENTRY) {
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_task_error");
- write_chart_dimension("calls", values[4]);
+ write_chart_dimension("calls", w->process.task_err);
ebpf_write_end_chart();
}
}
-
- ebpf_process_remove_pids();
}
/*****************************************************************
@@ -287,34 +221,20 @@ static void ebpf_read_process_hash_global_tables(netdata_idx_t *stats, int maps_
*
* @param maps_per_core do I need to read all cores?
*/
-static void ebpf_update_process_cgroup(int maps_per_core)
+static void ebpf_update_process_cgroup()
{
ebpf_cgroup_target_t *ect ;
- int pid_fd = process_maps[NETDATA_PROCESS_PID_TABLE].map_fd;
-
- size_t length = sizeof(ebpf_process_stat_t);
- if (maps_per_core)
- length *= ebpf_nprocs;
pthread_mutex_lock(&mutex_cgroup_shm);
for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
struct pid_on_target2 *pids;
for (pids = ect->pids; pids; pids = pids->next) {
int pid = pids->pid;
ebpf_process_stat_t *out = &pids->ps;
- if (global_process_stats[pid]) {
- ebpf_process_stat_t *in = global_process_stats[pid];
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ ebpf_process_stat_t *in = &local_pid->process;
memcpy(out, in, sizeof(ebpf_process_stat_t));
- } else {
- if (bpf_map_lookup_elem(pid_fd, &pid, process_stat_vector)) {
- memset(out, 0, sizeof(ebpf_process_stat_t));
- }
-
- ebpf_process_apps_accumulator(process_stat_vector, maps_per_core);
-
- memcpy(out, process_stat_vector, sizeof(ebpf_process_stat_t));
-
- memset(process_stat_vector, 0, length);
}
}
}
@@ -439,7 +359,7 @@ void ebpf_process_create_apps_charts(struct ebpf_module *em, void *ptr)
NETDATA_EBPF_MODULE_NAME_PROCESS);
ebpf_create_chart_labels("app_group", w->name, 1);
ebpf_commit_label();
- fprintf(stdout, "DIMENSION calls '' %s 1 1\n", ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
+ fprintf(stdout, "DIMENSION calls '' %s 1 1\n", ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX]);
ebpf_write_chart_cmd(NETDATA_APP_FAMILY,
w->clean_name,
@@ -454,7 +374,7 @@ void ebpf_process_create_apps_charts(struct ebpf_module *em, void *ptr)
NETDATA_EBPF_MODULE_NAME_PROCESS);
ebpf_create_chart_labels("app_group", w->name, 1);
ebpf_commit_label();
- fprintf(stdout, "DIMENSION calls '' %s 1 1\n", ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
+ fprintf(stdout, "DIMENSION calls '' %s 1 1\n", ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX]);
ebpf_write_chart_cmd(NETDATA_APP_FAMILY,
w->clean_name,
@@ -469,7 +389,7 @@ void ebpf_process_create_apps_charts(struct ebpf_module *em, void *ptr)
NETDATA_EBPF_MODULE_NAME_PROCESS);
ebpf_create_chart_labels("app_group", w->name, 1);
ebpf_commit_label();
- fprintf(stdout, "DIMENSION calls '' %s 1 1\n", ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
+ fprintf(stdout, "DIMENSION calls '' %s 1 1\n", ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX]);
ebpf_write_chart_cmd(NETDATA_APP_FAMILY,
w->clean_name,
@@ -484,7 +404,7 @@ void ebpf_process_create_apps_charts(struct ebpf_module *em, void *ptr)
NETDATA_EBPF_MODULE_NAME_PROCESS);
ebpf_create_chart_labels("app_group", w->name, 1);
ebpf_commit_label();
- fprintf(stdout, "DIMENSION calls '' %s 1 1\n", ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
+ fprintf(stdout, "DIMENSION calls '' %s 1 1\n", ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX]);
if (em->mode < MODE_ENTRY) {
ebpf_write_chart_cmd(NETDATA_APP_FAMILY,
@@ -500,7 +420,7 @@ void ebpf_process_create_apps_charts(struct ebpf_module *em, void *ptr)
NETDATA_EBPF_MODULE_NAME_PROCESS);
ebpf_create_chart_labels("app_group", w->name, 1);
ebpf_commit_label();
- fprintf(stdout, "DIMENSION calls '' %s 1 1\n", ebpf_algorithms[NETDATA_EBPF_ABSOLUTE_IDX]);
+ fprintf(stdout, "DIMENSION calls '' %s 1 1\n", ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX]);
}
w->charts_created |= 1<<EBPF_MODULE_PROCESS_IDX;
}
@@ -523,11 +443,11 @@ static void ebpf_obsolete_specific_process_charts(char *type, ebpf_module_t *em)
*
* @param em a pointer to `struct ebpf_module`
*/
-static void ebpf_obsolete_process_services(ebpf_module_t *em)
+static void ebpf_obsolete_process_services(ebpf_module_t *em, char *id)
{
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_TASK_PROCESS,
- "",
"Process started",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_PROCESS_GROUP,
@@ -537,8 +457,8 @@ static void ebpf_obsolete_process_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_TASK_THREAD,
- "",
"Threads started",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_PROCESS_GROUP,
@@ -548,8 +468,8 @@ static void ebpf_obsolete_process_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_TASK_CLOSE,
- "",
"Tasks starts exit process.",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_PROCESS_GROUP,
@@ -559,8 +479,8 @@ static void ebpf_obsolete_process_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_TASK_EXIT,
- "",
"Tasks closed",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_PROCESS_GROUP,
@@ -571,8 +491,8 @@ static void ebpf_obsolete_process_services(ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_TASK_ERROR,
- "",
"Errors to create process or threads.",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_PROCESS_GROUP,
@@ -593,12 +513,13 @@ static void ebpf_obsolete_process_services(ebpf_module_t *em)
static inline void ebpf_obsolete_process_cgroup_charts(ebpf_module_t *em) {
pthread_mutex_lock(&mutex_cgroup_shm);
- ebpf_obsolete_process_services(em);
-
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (ect->systemd)
+ if (ect->systemd) {
+ ebpf_obsolete_process_services(em, ect->name);
+
continue;
+ }
ebpf_obsolete_specific_process_charts(ect->name, em);
}
@@ -785,11 +706,6 @@ static void ebpf_process_exit(void *ptr)
ebpf_obsolete_process_global(em);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_process_stat)
- ebpf_statistic_obsolete_aral_chart(em, process_disable_priority);
-#endif
-
fflush(stdout);
pthread_mutex_unlock(&lock);
}
@@ -905,14 +821,14 @@ static void ebpf_send_specific_process_data(char *type, ebpf_process_stat_t *val
static void ebpf_create_specific_process_charts(char *type, ebpf_module_t *em)
{
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_TASK_PROCESS, "Process started",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_PROCESS_CGROUP_GROUP,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_PROCESS_GROUP,
NETDATA_CGROUP_PROCESS_CREATE_CONTEXT, NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5000,
ebpf_create_global_dimension, &process_publish_aggregated[NETDATA_KEY_PUBLISH_PROCESS_FORK],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_PROCESS);
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_TASK_THREAD, "Threads started",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_PROCESS_CGROUP_GROUP,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_PROCESS_GROUP,
NETDATA_CGROUP_THREAD_CREATE_CONTEXT, NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5001,
ebpf_create_global_dimension,
@@ -920,7 +836,7 @@ static void ebpf_create_specific_process_charts(char *type, ebpf_module_t *em)
1, em->update_every, NETDATA_EBPF_MODULE_NAME_PROCESS);
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_TASK_EXIT, "Tasks starts exit process.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_PROCESS_CGROUP_GROUP,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_PROCESS_GROUP,
NETDATA_CGROUP_PROCESS_EXIT_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5002,
ebpf_create_global_dimension,
@@ -928,7 +844,7 @@ static void ebpf_create_specific_process_charts(char *type, ebpf_module_t *em)
1, em->update_every, NETDATA_EBPF_MODULE_NAME_PROCESS);
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_TASK_CLOSE, "Tasks closed",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_PROCESS_CGROUP_GROUP,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_PROCESS_GROUP,
NETDATA_CGROUP_PROCESS_CLOSE_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5003,
ebpf_create_global_dimension,
@@ -937,7 +853,7 @@ static void ebpf_create_specific_process_charts(char *type, ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_TASK_ERROR, "Errors to create process or threads.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_PROCESS_CGROUP_GROUP,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_PROCESS_GROUP,
NETDATA_CGROUP_PROCESS_ERROR_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5004,
ebpf_create_global_dimension,
@@ -993,36 +909,98 @@ static void ebpf_obsolete_specific_process_charts(char *type, ebpf_module_t *em)
**/
static void ebpf_create_systemd_process_charts(ebpf_module_t *em)
{
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_TASK_PROCESS, "Process started",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_PROCESS_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20065,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_PROCESS_CREATE_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_PROCESS, em->update_every);
-
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_TASK_THREAD, "Threads started",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_PROCESS_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20066,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_THREAD_CREATE_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_PROCESS, em->update_every);
-
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_TASK_CLOSE, "Tasks starts exit process.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_PROCESS_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20067,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_PROCESS_EXIT_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_PROCESS, em->update_every);
-
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_TASK_EXIT, "Tasks closed",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_PROCESS_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20068,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_PROCESS_CLOSE_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_PROCESS, em->update_every);
+ static ebpf_systemd_args_t data_process = {
+ .title = "Process started",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_PROCESS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20065,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_PROCESS_CREATE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_PROCESS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_TASK_PROCESS,
+ .dimension = "calls"
+ };
- if (em->mode < MODE_ENTRY) {
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_TASK_ERROR, "Errors to create process or threads.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_APPS_PROCESS_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20069,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_PROCESS_ERROR_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_PROCESS, em->update_every);
+ static ebpf_systemd_args_t data_thread = {
+ .title = "Threads started",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_PROCESS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20066,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_THREAD_CREATE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_PROCESS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_TASK_THREAD,
+ .dimension = "calls"
+ };
+
+ static ebpf_systemd_args_t task_exit = {
+ .title = "Tasks starts exit process.",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_PROCESS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20067,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_PROCESS_EXIT_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_PROCESS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_TASK_CLOSE,
+ .dimension = "calls"
+ };
+
+ static ebpf_systemd_args_t task_closed = {
+ .title = "Tasks closed",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_PROCESS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20068,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_PROCESS_CLOSE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_PROCESS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_TASK_EXIT,
+ .dimension = "calls"
+ };
+
+ static ebpf_systemd_args_t task_error = {
+ .title = "Errors to create process or threads.",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_PROCESS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20069,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_PROCESS_ERROR_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_PROCESS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_TASK_ERROR,
+ .dimension = "calls"
+ };
+
+ ebpf_cgroup_target_t *w;
+ netdata_run_mode_t mode = em->mode;
+ if (!task_exit.update_every)
+ data_process.update_every = data_thread.update_every = task_exit.update_every =
+ task_closed.update_every = task_error.update_every = em->update_every;
+
+ for (w = ebpf_cgroup_pids; w; w = w->next) {
+ if (unlikely(!w->systemd || w->flags & NETDATA_EBPF_SERVICES_HAS_PROCESS_CHART))
+ continue;
+
+ data_process.id = data_thread.id = task_exit.id = task_closed.id = task_error.id = w->name;
+ ebpf_create_charts_on_systemd(&data_process);
+
+ ebpf_create_charts_on_systemd(&data_thread);
+
+ ebpf_create_charts_on_systemd(&task_exit);
+
+ ebpf_create_charts_on_systemd(&task_closed);
+ if (mode < MODE_ENTRY) {
+ ebpf_create_charts_on_systemd(&task_error);
+ }
+ w->flags |= NETDATA_EBPF_SERVICES_HAS_PROCESS_CHART;
}
}
@@ -1036,46 +1014,32 @@ static void ebpf_create_systemd_process_charts(ebpf_module_t *em)
static void ebpf_send_systemd_process_charts(ebpf_module_t *em)
{
ebpf_cgroup_target_t *ect;
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_TASK_PROCESS, "");
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_ps.create_process);
+ if (unlikely(!(ect->flags & NETDATA_EBPF_SERVICES_HAS_PROCESS_CHART)) ) {
+ continue;
}
- }
- ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_TASK_THREAD, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_ps.create_thread);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_TASK_PROCESS);
+ write_chart_dimension("calls", ect->publish_systemd_ps.create_process);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_TASK_EXIT, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_ps.exit_call);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_TASK_THREAD);
+ write_chart_dimension("calls", ect->publish_systemd_ps.create_thread);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_TASK_CLOSE, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_ps.release_call);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_TASK_EXIT);
+ write_chart_dimension("calls", ect->publish_systemd_ps.exit_call);
+ ebpf_write_end_chart();
- if (em->mode < MODE_ENTRY) {
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_TASK_ERROR, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_ps.task_err);
- }
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_TASK_CLOSE);
+ write_chart_dimension("calls", ect->publish_systemd_ps.release_call);
ebpf_write_end_chart();
+
+ if (em->mode < MODE_ENTRY) {
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_TASK_ERROR);
+ write_chart_dimension("calls", ect->publish_systemd_ps.task_err);
+ ebpf_write_end_chart();
+ }
}
}
@@ -1086,18 +1050,13 @@ static void ebpf_send_systemd_process_charts(ebpf_module_t *em)
*/
static void ebpf_process_send_cgroup_data(ebpf_module_t *em)
{
- if (!ebpf_cgroup_pids)
- return;
-
pthread_mutex_lock(&mutex_cgroup_shm);
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
ebpf_process_sum_cgroup_pids(&ect->publish_systemd_ps, ect->pids);
}
- int has_systemd = shm_ebpf_cgroup.header->systemd_enabled;
-
- if (has_systemd) {
+ if (shm_ebpf_cgroup.header->systemd_enabled) {
if (send_cgroup_chart) {
ebpf_create_systemd_process_charts(em);
}
@@ -1181,7 +1140,7 @@ static void process_collector(ebpf_module_t *em)
if (ebpf_all_pids_count > 0) {
if (cgroups && shm_ebpf_cgroup.header) {
- ebpf_update_process_cgroup(maps_per_core);
+ ebpf_update_process_cgroup();
}
}
@@ -1195,11 +1154,6 @@ static void process_collector(ebpf_module_t *em)
ebpf_process_send_apps_data(apps_groups_root_target, em);
}
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_process_stat)
- ebpf_send_data_aral_chart(ebpf_aral_process_stat, em);
-#endif
-
if (cgroups && shm_ebpf_cgroup.header) {
ebpf_process_send_cgroup_data(em);
}
@@ -1240,8 +1194,6 @@ static void ebpf_process_allocate_global_vectors(size_t length)
memset(process_publish_aggregated, 0, length * sizeof(netdata_publish_syscall_t));
process_hash_values = callocz(ebpf_nprocs, sizeof(netdata_idx_t));
process_stat_vector = callocz(ebpf_nprocs, sizeof(ebpf_process_stat_t));
-
- global_process_stats = callocz((size_t)pid_max, sizeof(ebpf_process_stat_t *));
}
static void change_syscalls()
@@ -1351,11 +1303,6 @@ void *ebpf_process_thread(void *ptr)
ebpf_update_stats(&plugin_statistics, em);
ebpf_update_kernel_memory_with_vector(&plugin_statistics, em->maps, EBPF_ACTION_STAT_ADD);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_process_stat)
- process_disable_priority = ebpf_statistic_create_aral_chart(NETDATA_EBPF_PROC_ARAL_NAME, em);
-#endif
-
pthread_mutex_unlock(&lock);
process_collector(em);
diff --git a/collectors/ebpf.plugin/ebpf_process.h b/src/collectors/ebpf.plugin/ebpf_process.h
index 310b321d6..18ffec1ff 100644
--- a/collectors/ebpf.plugin/ebpf_process.h
+++ b/src/collectors/ebpf.plugin/ebpf_process.h
@@ -9,7 +9,6 @@
// Groups used on Dashboard
#define NETDATA_PROCESS_GROUP "processes"
-#define NETDATA_PROCESS_CGROUP_GROUP "processes (eBPF)"
// Global chart name
#define NETDATA_EXIT_SYSCALL "exit"
@@ -34,11 +33,11 @@
#define NETDATA_CGROUP_PROCESS_EXIT_CONTEXT "cgroup.task_exit"
#define NETDATA_CGROUP_PROCESS_ERROR_CONTEXT "cgroup.task_error"
-#define NETDATA_SYSTEMD_PROCESS_CREATE_CONTEXT "services.process_create"
-#define NETDATA_SYSTEMD_THREAD_CREATE_CONTEXT "services.thread_create"
-#define NETDATA_SYSTEMD_PROCESS_CLOSE_CONTEXT "services.task_close"
-#define NETDATA_SYSTEMD_PROCESS_EXIT_CONTEXT "services.task_exit"
-#define NETDATA_SYSTEMD_PROCESS_ERROR_CONTEXT "services.task_error"
+#define NETDATA_SYSTEMD_PROCESS_CREATE_CONTEXT "systemd.services.process_create"
+#define NETDATA_SYSTEMD_THREAD_CREATE_CONTEXT "systemd.services.thread_create"
+#define NETDATA_SYSTEMD_PROCESS_CLOSE_CONTEXT "systemd.services.task_close"
+#define NETDATA_SYSTEMD_PROCESS_EXIT_CONTEXT "systemd.services.task_exit"
+#define NETDATA_SYSTEMD_PROCESS_ERROR_CONTEXT "systemd.services.task_error"
#define NETDATA_EBPF_CGROUP_UPDATE 30
diff --git a/collectors/ebpf.plugin/ebpf_shm.c b/src/collectors/ebpf.plugin/ebpf_shm.c
index f14eb67d0..816e68cfd 100644
--- a/collectors/ebpf.plugin/ebpf_shm.c
+++ b/src/collectors/ebpf.plugin/ebpf_shm.c
@@ -54,6 +54,17 @@ netdata_ebpf_targets_t shm_targets[] = { {.name = "shmget", .mode = EBPF_LOAD_TR
int shm_disable_priority;
#endif
+struct netdata_static_thread ebpf_read_shm = {
+ .name = "EBPF_READ_SHM",
+ .config_section = NULL,
+ .config_name = NULL,
+ .env_name = NULL,
+ .enabled = 1,
+ .thread = NULL,
+ .init_routine = NULL,
+ .start_routine = NULL
+};
+
#ifdef LIBBPF_MAJOR_VERSION
/*****************************************************************
*
@@ -89,7 +100,6 @@ static void ebpf_disable_probe(struct shm_bpf *obj)
bpf_program__set_autoload(obj->progs.netdata_shmat_probe, false);
bpf_program__set_autoload(obj->progs.netdata_shmdt_probe, false);
bpf_program__set_autoload(obj->progs.netdata_shmctl_probe, false);
- bpf_program__set_autoload(obj->progs.netdata_shm_release_task_probe, false);
}
/*
@@ -105,7 +115,6 @@ static void ebpf_disable_trampoline(struct shm_bpf *obj)
bpf_program__set_autoload(obj->progs.netdata_shmat_fentry, false);
bpf_program__set_autoload(obj->progs.netdata_shmdt_fentry, false);
bpf_program__set_autoload(obj->progs.netdata_shmctl_fentry, false);
- bpf_program__set_autoload(obj->progs.netdata_shm_release_task_fentry, false);
}
/**
@@ -138,9 +147,6 @@ static void ebpf_set_trampoline_target(struct shm_bpf *obj)
shm_targets[NETDATA_KEY_SHMCTL_CALL].name, running_on_kernel);
bpf_program__set_attach_target(obj->progs.netdata_shmctl_fentry, 0,
syscall);
-
- bpf_program__set_attach_target(obj->progs.netdata_shm_release_task_fentry, 0,
- EBPF_COMMON_FNCT_CLEAN_UP);
}
/**
@@ -160,7 +166,7 @@ static int ebpf_shm_attach_probe(struct shm_bpf *obj)
obj->links.netdata_shmget_probe = bpf_program__attach_kprobe(obj->progs.netdata_shmget_probe,
false, syscall);
- int ret = (int)libbpf_get_error(obj->links.netdata_shmget_probe);
+ long ret = libbpf_get_error(obj->links.netdata_shmget_probe);
if (ret)
return -1;
@@ -168,7 +174,7 @@ static int ebpf_shm_attach_probe(struct shm_bpf *obj)
shm_targets[NETDATA_KEY_SHMAT_CALL].name, running_on_kernel);
obj->links.netdata_shmat_probe = bpf_program__attach_kprobe(obj->progs.netdata_shmat_probe,
false, syscall);
- ret = (int)libbpf_get_error(obj->links.netdata_shmat_probe);
+ ret = libbpf_get_error(obj->links.netdata_shmat_probe);
if (ret)
return -1;
@@ -176,7 +182,7 @@ static int ebpf_shm_attach_probe(struct shm_bpf *obj)
shm_targets[NETDATA_KEY_SHMDT_CALL].name, running_on_kernel);
obj->links.netdata_shmdt_probe = bpf_program__attach_kprobe(obj->progs.netdata_shmdt_probe,
false, syscall);
- ret = (int)libbpf_get_error(obj->links.netdata_shmdt_probe);
+ ret = libbpf_get_error(obj->links.netdata_shmdt_probe);
if (ret)
return -1;
@@ -184,17 +190,10 @@ static int ebpf_shm_attach_probe(struct shm_bpf *obj)
shm_targets[NETDATA_KEY_SHMCTL_CALL].name, running_on_kernel);
obj->links.netdata_shmctl_probe = bpf_program__attach_kprobe(obj->progs.netdata_shmctl_probe,
false, syscall);
- ret = (int)libbpf_get_error(obj->links.netdata_shmctl_probe);
- if (ret)
- return -1;
-
- obj->links.netdata_shm_release_task_probe = bpf_program__attach_kprobe(obj->progs.netdata_shm_release_task_probe,
- false, EBPF_COMMON_FNCT_CLEAN_UP);
- ret = (int)libbpf_get_error(obj->links.netdata_shm_release_task_probe);
+ ret = libbpf_get_error(obj->links.netdata_shmctl_probe);
if (ret)
return -1;
-
return 0;
}
@@ -211,19 +210,6 @@ static void ebpf_shm_set_hash_tables(struct shm_bpf *obj)
}
/**
- * Disable Release Task
- *
- * Disable release task when apps is not enabled.
- *
- * @param obj is the main structure for bpf objects.
- */
-static void ebpf_shm_disable_release_task(struct shm_bpf *obj)
-{
- bpf_program__set_autoload(obj->progs.netdata_shm_release_task_probe, false);
- bpf_program__set_autoload(obj->progs.netdata_shm_release_task_fentry, false);
-}
-
-/**
* Adjust Map Size
*
* Resize maps according input from users.
@@ -271,8 +257,6 @@ static inline int ebpf_shm_load_and_attach(struct shm_bpf *obj, ebpf_module_t *e
}
ebpf_shm_adjust_map(obj, em);
- if (!em->apps_charts && !em->cgroup_charts)
- ebpf_shm_disable_release_task(obj);
int ret = shm_bpf__load(obj);
if (!ret) {
@@ -301,11 +285,11 @@ static void ebpf_obsolete_specific_shm_charts(char *type, int update_every);
*
* @param em a pointer to `struct ebpf_module`
*/
-static void ebpf_obsolete_shm_services(ebpf_module_t *em)
+static void ebpf_obsolete_shm_services(ebpf_module_t *em, char *id)
{
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SHMGET_CHART,
- "",
"Calls to syscall shmget(2).",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_IPC_SHM_GROUP,
@@ -315,8 +299,8 @@ static void ebpf_obsolete_shm_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SHMAT_CHART,
- "",
"Calls to syscall shmat(2).",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_IPC_SHM_GROUP,
@@ -326,8 +310,8 @@ static void ebpf_obsolete_shm_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SHMDT_CHART,
- "",
"Calls to syscall shmdt(2).",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_IPC_SHM_GROUP,
@@ -337,8 +321,8 @@ static void ebpf_obsolete_shm_services(ebpf_module_t *em)
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SHMCTL_CHART,
- "",
"Calls to syscall shmctl(2).",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_IPC_SHM_GROUP,
@@ -358,12 +342,13 @@ static void ebpf_obsolete_shm_services(ebpf_module_t *em)
static inline void ebpf_obsolete_shm_cgroup_charts(ebpf_module_t *em) {
pthread_mutex_lock(&mutex_cgroup_shm);
- ebpf_obsolete_shm_services(em);
-
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (ect->systemd)
+ if (ect->systemd) {
+ ebpf_obsolete_shm_services(em, ect->name);
+
continue;
+ }
ebpf_obsolete_specific_shm_charts(ect->name, em->update_every);
}
@@ -381,6 +366,7 @@ void ebpf_obsolete_shm_apps_charts(struct ebpf_module *em)
{
struct ebpf_target *w;
int update_every = em->update_every;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = apps_groups_root_target; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_SHM_IDX))))
continue;
@@ -431,6 +417,7 @@ void ebpf_obsolete_shm_apps_charts(struct ebpf_module *em)
w->charts_created &= ~(1<<EBPF_MODULE_SHM_IDX);
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -465,6 +452,9 @@ static void ebpf_shm_exit(void *ptr)
{
ebpf_module_t *em = (ebpf_module_t *)ptr;
+ if (ebpf_read_shm.thread)
+ netdata_thread_cancel(*ebpf_read_shm.thread);
+
if (em->enabled == NETDATA_THREAD_EBPF_FUNCTION_RUNNING) {
pthread_mutex_lock(&lock);
if (em->cgroup_charts) {
@@ -478,11 +468,6 @@ static void ebpf_shm_exit(void *ptr)
ebpf_obsolete_shm_global(em);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_shm_pid)
- ebpf_statistic_obsolete_aral_chart(em, shm_disable_priority);
-#endif
-
fflush(stdout);
pthread_mutex_unlock(&lock);
}
@@ -534,38 +519,16 @@ static void shm_apps_accumulator(netdata_publish_shm_t *out, int maps_per_core)
}
/**
- * Fill PID
- *
- * Fill PID structures
- *
- * @param current_pid pid that we are collecting data
- * @param out values read from hash tables;
- */
-static void shm_fill_pid(uint32_t current_pid, netdata_publish_shm_t *publish)
-{
- netdata_publish_shm_t *curr = shm_pid[current_pid];
- if (!curr) {
- curr = ebpf_shm_stat_get( );
- shm_pid[current_pid] = curr;
- }
-
- memcpy(curr, publish, sizeof(netdata_publish_shm_t));
-}
-
-/**
* Update cgroup
*
* Update cgroup data based in
*
* @param maps_per_core do I need to read all cores?
*/
-static void ebpf_update_shm_cgroup(int maps_per_core)
+static void ebpf_update_shm_cgroup()
{
netdata_publish_shm_t *cv = shm_vector;
- int fd = shm_maps[NETDATA_PID_SHM_TABLE].map_fd;
size_t length = sizeof(netdata_publish_shm_t);
- if (maps_per_core)
- length *= ebpf_nprocs;
ebpf_cgroup_target_t *ect;
@@ -577,20 +540,11 @@ static void ebpf_update_shm_cgroup(int maps_per_core)
for (pids = ect->pids; pids; pids = pids->next) {
int pid = pids->pid;
netdata_publish_shm_t *out = &pids->shm;
- if (likely(shm_pid) && shm_pid[pid]) {
- netdata_publish_shm_t *in = shm_pid[pid];
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ netdata_publish_shm_t *in = &local_pid->shm;
memcpy(out, in, sizeof(netdata_publish_shm_t));
- } else {
- if (!bpf_map_lookup_elem(fd, &pid, cv)) {
- shm_apps_accumulator(cv, maps_per_core);
-
- memcpy(out, cv, sizeof(netdata_publish_shm_t));
-
- // now that we've consumed the value, zero it out in the map.
- memset(cv, 0, length);
- bpf_map_update_elem(fd, &pid, cv, BPF_EXIST);
- }
}
}
}
@@ -604,33 +558,42 @@ static void ebpf_update_shm_cgroup(int maps_per_core)
*
* @param maps_per_core do I need to read all cores?
*/
-static void read_shm_apps_table(int maps_per_core)
+static void ebpf_read_shm_apps_table(int maps_per_core, int max_period)
{
netdata_publish_shm_t *cv = shm_vector;
- uint32_t key;
- struct ebpf_pid_stat *pids = ebpf_root_of_pids;
int fd = shm_maps[NETDATA_PID_SHM_TABLE].map_fd;
size_t length = sizeof(netdata_publish_shm_t);
if (maps_per_core)
length *= ebpf_nprocs;
- while (pids) {
- key = pids->pid;
-
+ uint32_t key = 0, next_key = 0;
+ while (bpf_map_get_next_key(fd, &key, &next_key) == 0) {
if (bpf_map_lookup_elem(fd, &key, cv)) {
- pids = pids->next;
- continue;
+ goto end_shm_loop;
}
shm_apps_accumulator(cv, maps_per_core);
- shm_fill_pid(key, cv);
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(key, 0);
+ if (!local_pid)
+ goto end_shm_loop;
+
+ netdata_publish_shm_t *publish = &local_pid->shm;
+ if (!publish->ct || publish->ct != cv->ct) {
+ memcpy(publish, &cv[0], sizeof(netdata_publish_shm_t));
+ local_pid->not_updated = 0;
+ } else if (++local_pid->not_updated >= max_period){
+ bpf_map_delete_elem(fd, &key);
+ local_pid->not_updated = 0;
+ }
+
+end_shm_loop:
// now that we've consumed the value, zero it out in the map.
memset(cv, 0, length);
bpf_map_update_elem(fd, &key, cv, BPF_EXIST);
- pids = pids->next;
+ key = next_key;
}
}
@@ -689,10 +652,12 @@ static void ebpf_shm_read_global_table(netdata_idx_t *stats, int maps_per_core)
*/
static void ebpf_shm_sum_pids(netdata_publish_shm_t *shm, struct ebpf_pid_on_target *root)
{
+ memset(shm, 0, sizeof(netdata_publish_shm_t));
while (root) {
int32_t pid = root->pid;
- netdata_publish_shm_t *w = shm_pid[pid];
- if (w) {
+ ebpf_pid_stat_t *pid_stat = ebpf_get_pid_entry(pid, 0);
+ if (pid_stat) {
+ netdata_publish_shm_t *w = &pid_stat->shm;
shm->get += w->get;
shm->at += w->at;
shm->dt += w->dt;
@@ -716,12 +681,11 @@ static void ebpf_shm_sum_pids(netdata_publish_shm_t *shm, struct ebpf_pid_on_tar
void ebpf_shm_send_apps_data(struct ebpf_target *root)
{
struct ebpf_target *w;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = root; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_SHM_IDX))))
continue;
- ebpf_shm_sum_pids(&w->shm, w->root_pid);
-
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_shmget_call");
write_chart_dimension("calls", (long long) w->shm.get);
ebpf_write_end_chart();
@@ -738,6 +702,7 @@ void ebpf_shm_send_apps_data(struct ebpf_target *root)
write_chart_dimension("calls", (long long) w->shm.ctl);
ebpf_write_end_chart();
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -875,41 +840,82 @@ static void ebpf_obsolete_specific_shm_charts(char *type, int update_every)
**/
static void ebpf_create_systemd_shm_charts(int update_every)
{
- ebpf_create_charts_on_systemd(NETDATA_SHMGET_CHART,
- "Calls to syscall shmget(2).",
- EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_IPC_SHM_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- 20191,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SYSTEMD_SHM_GET_CONTEXT, NETDATA_EBPF_MODULE_NAME_SHM, update_every);
+ static ebpf_systemd_args_t data_shmget = {
+ .title = "Calls to syscall shmget(2).",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_IPC_SHM_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20191,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_SHM_GET_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SHM,
+ .update_every = 0,
+ .suffix = NETDATA_SHMGET_CHART,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_SHMAT_CHART,
- "Calls to syscall shmat(2).",
- EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_IPC_SHM_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- 20192,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SYSTEMD_SHM_AT_CONTEXT, NETDATA_EBPF_MODULE_NAME_SHM, update_every);
+ static ebpf_systemd_args_t data_shmat = {
+ .title = "Calls to syscall shmat(2).",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_IPC_SHM_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20192,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_SHM_AT_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SHM,
+ .update_every = 0,
+ .suffix = NETDATA_SHMAT_CHART,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_SHMDT_CHART,
- "Calls to syscall shmdt(2).",
- EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_IPC_SHM_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- 20193,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SYSTEMD_SHM_DT_CONTEXT, NETDATA_EBPF_MODULE_NAME_SHM, update_every);
+ static ebpf_systemd_args_t data_shmdt = {
+ .title = "Calls to syscall shmdt(2).",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_IPC_SHM_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20193,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_SHM_DT_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SHM,
+ .update_every = 0,
+ .suffix = NETDATA_SHMDT_CHART,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_SHMCTL_CHART,
- "Calls to syscall shmctl(2).",
- EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_IPC_SHM_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- 20193,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SYSTEMD_SHM_CTL_CONTEXT, NETDATA_EBPF_MODULE_NAME_SHM, update_every);
+ static ebpf_systemd_args_t data_shmctl = {
+ .title = "Calls to syscall shmctl(2).",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_IPC_SHM_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20194,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_SHM_CTL_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SHM,
+ .update_every = 0,
+ .suffix = NETDATA_SHMCTL_CHART,
+ .dimension = "calls"
+ };
+
+ if (!data_shmget.update_every)
+ data_shmat.update_every = data_shmctl.update_every =
+ data_shmdt.update_every = data_shmget.update_every = update_every;
+
+ ebpf_cgroup_target_t *w;
+ for (w = ebpf_cgroup_pids; w; w = w->next) {
+ if (unlikely(!w->systemd || w->flags & NETDATA_EBPF_SERVICES_HAS_SHM_CHART))
+ continue;
+
+ data_shmat.id = data_shmctl.id = data_shmdt.id = data_shmget.id = w->name;
+ ebpf_create_charts_on_systemd(&data_shmat);
+
+ ebpf_create_charts_on_systemd(&data_shmctl);
+
+ ebpf_create_charts_on_systemd(&data_shmdt);
+
+ ebpf_create_charts_on_systemd(&data_shmget);
+
+ w->flags |= NETDATA_EBPF_SERVICES_HAS_SHM_CHART;
+ }
}
/**
@@ -920,37 +926,27 @@ static void ebpf_create_systemd_shm_charts(int update_every)
static void ebpf_send_systemd_shm_charts()
{
ebpf_cgroup_target_t *ect;
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SHMGET_CHART, "");
for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_shm.get);
+ if (unlikely(!(ect->flags & NETDATA_EBPF_SERVICES_HAS_SHM_CHART)) ) {
+ continue;
}
- }
- ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SHMAT_CHART, "");
- for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_shm.at);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SHMGET_CHART);
+ write_chart_dimension("calls", (long long)ect->publish_shm.get);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SHMDT_CHART, "");
- for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_shm.dt);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SHMAT_CHART);
+ write_chart_dimension("calls", (long long)ect->publish_shm.at);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SHMCTL_CHART, "");
- for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_shm.ctl);
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SHMDT_CHART);
+ write_chart_dimension("calls", (long long)ect->publish_shm.dt);
+ ebpf_write_end_chart();
+
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SHMCTL_CHART);
+ write_chart_dimension("calls", (long long)ect->publish_shm.ctl);
+ ebpf_write_end_chart();
}
- ebpf_write_end_chart();
}
/*
@@ -987,17 +983,13 @@ static void ebpf_send_specific_shm_data(char *type, netdata_publish_shm_t *value
*/
void ebpf_shm_send_cgroup_data(int update_every)
{
- if (!ebpf_cgroup_pids)
- return;
-
pthread_mutex_lock(&mutex_cgroup_shm);
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
ebpf_shm_sum_cgroup_pids(&ect->publish_shm, ect->pids);
}
- int has_systemd = shm_ebpf_cgroup.header->systemd_enabled;
- if (has_systemd) {
+ if (shm_ebpf_cgroup.header->systemd_enabled) {
if (send_cgroup_chart) {
ebpf_create_systemd_shm_charts(update_every);
}
@@ -1028,6 +1020,72 @@ void ebpf_shm_send_cgroup_data(int update_every)
}
/**
+ * Resume apps data
+ */
+void ebpf_shm_resume_apps_data() {
+ struct ebpf_target *w;
+ for (w = apps_groups_root_target; w; w = w->next) {
+ if (unlikely(!(w->charts_created & (1 << EBPF_MODULE_SHM_IDX))))
+ continue;
+
+ ebpf_shm_sum_pids(&w->shm, w->root_pid);
+ }
+}
+
+/**
+ * DCstat thread
+ *
+ * Thread used to generate dcstat charts.
+ *
+ * @param ptr a pointer to `struct ebpf_module`
+ *
+ * @return It always return NULL
+ */
+void *ebpf_read_shm_thread(void *ptr)
+{
+ heartbeat_t hb;
+ heartbeat_init(&hb);
+
+ ebpf_module_t *em = (ebpf_module_t *)ptr;
+
+ int maps_per_core = em->maps_per_core;
+ int update_every = em->update_every;
+
+ int counter = update_every - 1;
+
+ uint32_t lifetime = em->lifetime;
+ uint32_t running_time = 0;
+ usec_t period = update_every * USEC_PER_SEC;
+ int max_period = update_every * EBPF_CLEANUP_FACTOR;
+ while (!ebpf_plugin_exit && running_time < lifetime) {
+ (void)heartbeat_next(&hb, period);
+ if (ebpf_plugin_exit || ++counter != update_every)
+ continue;
+
+ netdata_thread_disable_cancelability();
+
+ pthread_mutex_lock(&collect_data_mutex);
+ ebpf_read_shm_apps_table(maps_per_core, max_period);
+ ebpf_shm_resume_apps_data();
+ pthread_mutex_unlock(&collect_data_mutex);
+
+ counter = 0;
+
+ pthread_mutex_lock(&ebpf_exit_cleanup);
+ if (running_time && !em->running_time)
+ running_time = update_every;
+ else
+ running_time += update_every;
+
+ em->running_time = running_time;
+ pthread_mutex_unlock(&ebpf_exit_cleanup);
+ netdata_thread_enable_cancelability();
+ }
+
+ return NULL;
+}
+
+/**
* Main loop for this collector.
*/
static void shm_collector(ebpf_module_t *em)
@@ -1050,34 +1108,23 @@ static void shm_collector(ebpf_module_t *em)
counter = 0;
netdata_apps_integration_flags_t apps = em->apps_charts;
ebpf_shm_read_global_table(stats, maps_per_core);
- pthread_mutex_lock(&collect_data_mutex);
- if (apps) {
- read_shm_apps_table(maps_per_core);
- }
+ pthread_mutex_lock(&lock);
if (cgroups) {
- ebpf_update_shm_cgroup(maps_per_core);
+ ebpf_update_shm_cgroup();
}
- pthread_mutex_lock(&lock);
-
shm_send_global();
if (apps & NETDATA_EBPF_APPS_FLAG_CHART_CREATED) {
ebpf_shm_send_apps_data(apps_groups_root_target);
}
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_shm_pid)
- ebpf_send_data_aral_chart(ebpf_aral_shm_pid, em);
-#endif
-
if (cgroups) {
ebpf_shm_send_cgroup_data(update_every);
}
pthread_mutex_unlock(&lock);
- pthread_mutex_unlock(&collect_data_mutex);
pthread_mutex_lock(&ebpf_exit_cleanup);
if (running_time && !em->running_time)
@@ -1186,12 +1233,8 @@ void ebpf_shm_create_apps_charts(struct ebpf_module *em, void *ptr)
*/
static void ebpf_shm_allocate_global_vectors(int apps)
{
- if (apps) {
- ebpf_shm_aral_init();
- shm_pid = callocz((size_t)pid_max, sizeof(netdata_publish_shm_t *));
- shm_vector = callocz((size_t)ebpf_nprocs, sizeof(netdata_publish_shm_t));
- }
-
+ UNUSED(apps);
+ shm_vector = callocz((size_t)ebpf_nprocs, sizeof(netdata_publish_shm_t));
shm_values = callocz((size_t)ebpf_nprocs, sizeof(netdata_idx_t));
memset(shm_hash_values, 0, sizeof(shm_hash_values));
@@ -1310,13 +1353,15 @@ void *ebpf_shm_thread(void *ptr)
ebpf_create_shm_charts(em->update_every);
ebpf_update_stats(&plugin_statistics, em);
ebpf_update_kernel_memory_with_vector(&plugin_statistics, em->maps, EBPF_ACTION_STAT_ADD);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_shm_pid)
- shm_disable_priority = ebpf_statistic_create_aral_chart(NETDATA_EBPF_SHM_ARAL_NAME, em);
-#endif
-
pthread_mutex_unlock(&lock);
+ ebpf_read_shm.thread = mallocz(sizeof(netdata_thread_t));
+ netdata_thread_create(ebpf_read_shm.thread,
+ ebpf_read_shm.name,
+ NETDATA_THREAD_OPTION_DEFAULT,
+ ebpf_read_shm_thread,
+ em);
+
shm_collector(em);
endshm:
diff --git a/collectors/ebpf.plugin/ebpf_shm.h b/src/collectors/ebpf.plugin/ebpf_shm.h
index a415006e6..5a670b1b5 100644
--- a/collectors/ebpf.plugin/ebpf_shm.h
+++ b/src/collectors/ebpf.plugin/ebpf_shm.h
@@ -23,15 +23,15 @@
#define NETDATA_CGROUP_SHM_DT_CONTEXT "cgroup.shmdt"
#define NETDATA_CGROUP_SHM_CTL_CONTEXT "cgroup.shmctl"
-#define NETDATA_SYSTEMD_SHM_GET_CONTEXT "services.shmget"
-#define NETDATA_SYSTEMD_SHM_AT_CONTEXT "services.shmat"
-#define NETDATA_SYSTEMD_SHM_DT_CONTEXT "services.shmdt"
-#define NETDATA_SYSTEMD_SHM_CTL_CONTEXT "services.shmctl"
-
-// ARAL name
-#define NETDATA_EBPF_SHM_ARAL_NAME "ebpf_shm"
+#define NETDATA_SYSTEMD_SHM_GET_CONTEXT "systemd.services.shmget"
+#define NETDATA_SYSTEMD_SHM_AT_CONTEXT "systemd.services.shmat"
+#define NETDATA_SYSTEMD_SHM_DT_CONTEXT "systemd.services.shmdt"
+#define NETDATA_SYSTEMD_SHM_CTL_CONTEXT "systemd.services.shmctl"
typedef struct netdata_publish_shm {
+ uint64_t ct;
+ char name[TASK_COMM_LEN];
+
uint64_t get;
uint64_t at;
uint64_t dt;
diff --git a/collectors/ebpf.plugin/ebpf_socket.c b/src/collectors/ebpf.plugin/ebpf_socket.c
index bbb5dca1b..5361526df 100644
--- a/collectors/ebpf.plugin/ebpf_socket.c
+++ b/src/collectors/ebpf.plugin/ebpf_socket.c
@@ -107,10 +107,6 @@ struct netdata_static_thread ebpf_read_socket = {
ARAL *aral_socket_table = NULL;
-#ifdef NETDATA_DEV_MODE
-int socket_disable_priority;
-#endif
-
#ifdef LIBBPF_MAJOR_VERSION
/**
* Disable Probe
@@ -510,12 +506,12 @@ static void ebpf_socket_free(ebpf_module_t *em )
*
* @param update_every value to overwrite the update frequency set by the server.
**/
-static void ebpf_obsolete_systemd_socket_charts(int update_every)
+static void ebpf_obsolete_systemd_socket_charts(int update_every, char *id)
{
int order = 20080;
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_NET_APPS_CONNECTION_TCP_V4,
- "",
"Calls to tcp_v4_connection",
EBPF_COMMON_DIMENSION_CONNECTIONS,
NETDATA_APPS_NET_GROUP,
@@ -526,8 +522,8 @@ static void ebpf_obsolete_systemd_socket_charts(int update_every)
if (tcp_v6_connect_address.type == 'T') {
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_NET_APPS_CONNECTION_TCP_V6,
- "",
"Calls to tcp_v6_connection",
EBPF_COMMON_DIMENSION_CONNECTIONS,
NETDATA_APPS_NET_GROUP,
@@ -538,8 +534,8 @@ static void ebpf_obsolete_systemd_socket_charts(int update_every)
}
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_NET_APPS_BANDWIDTH_RECV,
- "",
"Bytes received",
EBPF_COMMON_DIMENSION_BITS,
NETDATA_APPS_NET_GROUP,
@@ -549,8 +545,8 @@ static void ebpf_obsolete_systemd_socket_charts(int update_every)
update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_NET_APPS_BANDWIDTH_SENT,
- "",
"Bytes sent",
EBPF_COMMON_DIMENSION_BITS,
NETDATA_APPS_NET_GROUP,
@@ -560,8 +556,8 @@ static void ebpf_obsolete_systemd_socket_charts(int update_every)
update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_NET_APPS_BANDWIDTH_TCP_RECV_CALLS,
- "",
"Calls to tcp_cleanup_rbuf.",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_NET_GROUP,
@@ -571,8 +567,8 @@ static void ebpf_obsolete_systemd_socket_charts(int update_every)
update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_NET_APPS_BANDWIDTH_TCP_SEND_CALLS,
- "",
"Calls to tcp_sendmsg.",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_NET_GROUP,
@@ -582,8 +578,8 @@ static void ebpf_obsolete_systemd_socket_charts(int update_every)
update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_NET_APPS_BANDWIDTH_TCP_RETRANSMIT,
- "",
"Calls to tcp_retransmit",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_NET_GROUP,
@@ -593,8 +589,8 @@ static void ebpf_obsolete_systemd_socket_charts(int update_every)
update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_NET_APPS_BANDWIDTH_UDP_SEND_CALLS,
- "",
"Calls to udp_sendmsg",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_NET_GROUP,
@@ -604,8 +600,8 @@ static void ebpf_obsolete_systemd_socket_charts(int update_every)
update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_NET_APPS_BANDWIDTH_UDP_RECV_CALLS,
- "",
"Calls to udp_recvmsg",
EBPF_COMMON_DIMENSION_CALL,
NETDATA_APPS_NET_GROUP,
@@ -626,12 +622,13 @@ static void ebpf_obsolete_specific_socket_charts(char *type, int update_every);
static inline void ebpf_obsolete_socket_cgroup_charts(ebpf_module_t *em) {
pthread_mutex_lock(&mutex_cgroup_shm);
- ebpf_obsolete_systemd_socket_charts(em->update_every);
-
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (ect->systemd)
+ if (ect->systemd) {
+ ebpf_obsolete_systemd_socket_charts(em->update_every, ect->name);
+
continue;
+ }
ebpf_obsolete_specific_socket_charts(ect->name, em->update_every);
}
@@ -650,6 +647,7 @@ void ebpf_socket_obsolete_apps_charts(struct ebpf_module *em)
int order = 20130;
struct ebpf_target *w;
int update_every = em->update_every;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = apps_groups_root_target; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_SOCKET_IDX))))
continue;
@@ -757,6 +755,7 @@ void ebpf_socket_obsolete_apps_charts(struct ebpf_module *em)
w->charts_created &= ~(1<<EBPF_MODULE_SOCKET_IDX);
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -904,10 +903,6 @@ static void ebpf_socket_exit(void *ptr)
ebpf_socket_obsolete_global_charts(em);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_socket_pid)
- ebpf_statistic_obsolete_aral_chart(em, socket_disable_priority);
-#endif
pthread_mutex_unlock(&lock);
}
@@ -1031,83 +1026,57 @@ static void ebpf_socket_send_data(ebpf_module_t *em)
/**
* Send data to Netdata calling auxiliary functions.
- *
- * @param em the structure with thread information
- * @param root the target list.
*/
-void ebpf_socket_send_apps_data(ebpf_module_t *em, struct ebpf_target *root)
+void ebpf_socket_send_apps_data()
{
- UNUSED(em);
-
struct ebpf_target *w;
- // This algorithm is improved in https://github.com/netdata/netdata/pull/16030
- collected_number values[9];
-
- for (w = root; w; w = w->next) {
+ pthread_mutex_lock(&collect_data_mutex);
+ for (w = apps_groups_root_target; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_SOCKET_IDX))))
continue;
- struct ebpf_pid_on_target *move = w->root_pid;
- // Simplify algorithm, but others will appear only in https://github.com/netdata/netdata/pull/16030
- memset(values, 0, sizeof(values));
- while (move) {
- int32_t pid = move->pid;
- ebpf_socket_publish_apps_t *ws = socket_bandwidth_curr[pid];
- if (ws) {
- values[0] += (collected_number) ws->call_tcp_v4_connection;
- values[1] += (collected_number) ws->call_tcp_v6_connection;
- values[2] += (collected_number) ws->bytes_sent;
- values[3] += (collected_number) ws->bytes_received;
- values[4] += (collected_number) ws->call_tcp_sent;
- values[5] += (collected_number) ws->call_tcp_received;
- values[6] += (collected_number) ws->retransmit;
- values[7] += (collected_number) ws->call_udp_sent;
- values[8] += (collected_number) ws->call_udp_received;
- }
-
- move = move->next;
- }
-
+ ebpf_socket_publish_apps_t *values = &w->socket;
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_call_tcp_v4_connection");
- write_chart_dimension("connections", values[0]);
+ write_chart_dimension("connections", (collected_number) values->call_tcp_v4_connection);
ebpf_write_end_chart();
if (tcp_v6_connect_address.type == 'T') {
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_call_tcp_v6_connection");
- write_chart_dimension("calls", values[1]);
+ write_chart_dimension("calls", (collected_number) values->call_tcp_v6_connection);
ebpf_write_end_chart();
}
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_sock_bytes_sent");
// We multiply by 0.008, because we read bytes, but we display bits
- write_chart_dimension("bandwidth", ((values[2])*8)/1000);
+ write_chart_dimension("bandwidth", (collected_number) ((values->bytes_sent)*8)/1000);
ebpf_write_end_chart();
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_sock_bytes_received");
// We multiply by 0.008, because we read bytes, but we display bits
- write_chart_dimension("bandwidth", ((values[3])*8)/1000);
+ write_chart_dimension("bandwidth", (collected_number) ((values->bytes_received)*8)/1000);
ebpf_write_end_chart();
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_call_tcp_sendmsg");
- write_chart_dimension("calls", values[4]);
+ write_chart_dimension("calls", (collected_number) values->call_tcp_sent);
ebpf_write_end_chart();
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_call_tcp_cleanup_rbuf");
- write_chart_dimension("calls", values[5]);
+ write_chart_dimension("calls", (collected_number) values->call_tcp_received);
ebpf_write_end_chart();
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_call_tcp_retransmit");
- write_chart_dimension("calls", values[6]);
+ write_chart_dimension("calls", (collected_number) values->retransmit);
ebpf_write_end_chart();
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_call_udp_sendmsg");
- write_chart_dimension("calls", values[7]);
+ write_chart_dimension("calls", (collected_number) values->call_udp_sent);
ebpf_write_end_chart();
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_call_udp_recvmsg");
- write_chart_dimension("calls", values[8]);
+ write_chart_dimension("calls", (collected_number) values->call_udp_received);
ebpf_write_end_chart();
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/*****************************************************************
@@ -1809,6 +1778,41 @@ end_socket_loop:
}
netdata_thread_enable_cancelability();
}
+/**
+ * Resume apps data
+ */
+void ebpf_socket_resume_apps_data()
+{
+ struct ebpf_target *w;
+
+ for (w = apps_groups_root_target; w; w = w->next) {
+ if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_SOCKET_IDX))))
+ continue;
+
+ struct ebpf_pid_on_target *move = w->root_pid;
+
+ ebpf_socket_publish_apps_t *values = &w->socket;
+ memset(&w->socket, 0, sizeof(ebpf_socket_publish_apps_t));
+ while (move) {
+ int32_t pid = move->pid;
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ ebpf_socket_publish_apps_t *ws = &local_pid->socket;
+ values->call_tcp_v4_connection = ws->call_tcp_v4_connection;
+ values->call_tcp_v6_connection = ws->call_tcp_v6_connection;
+ values->bytes_sent = ws->bytes_sent;
+ values->bytes_received = ws->bytes_received;
+ values->call_tcp_sent = ws->call_tcp_sent;
+ values->call_tcp_received = ws->call_tcp_received;
+ values->retransmit = ws->retransmit;
+ values->call_udp_sent = ws->call_udp_sent;
+ values->call_udp_received = ws->call_udp_received;
+ }
+
+ move = move->next;
+ }
+ }
+}
/**
* Socket thread
@@ -1839,7 +1843,10 @@ void *ebpf_read_socket_thread(void *ptr)
if (ebpf_plugin_exit || ++counter != update_every)
continue;
+ pthread_mutex_lock(&collect_data_mutex);
ebpf_update_array_vectors(em);
+ ebpf_socket_resume_apps_data();
+ pthread_mutex_unlock(&collect_data_mutex);
counter = 0;
}
@@ -1992,23 +1999,23 @@ static void ebpf_socket_read_hash_global_tables(netdata_idx_t *stats, int maps_p
*/
void ebpf_socket_fill_publish_apps(uint32_t current_pid, netdata_socket_t *ns)
{
- ebpf_socket_publish_apps_t *curr = socket_bandwidth_curr[current_pid];
- if (!curr) {
- curr = ebpf_socket_stat_get();
- socket_bandwidth_curr[current_pid] = curr;
- }
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(current_pid, 0);
+ if (!local_pid)
+ return;
- curr->bytes_sent += ns->tcp.tcp_bytes_sent;
- curr->bytes_received += ns->tcp.tcp_bytes_received;
- curr->call_tcp_sent += ns->tcp.call_tcp_sent;
- curr->call_tcp_received += ns->tcp.call_tcp_received;
- curr->retransmit += ns->tcp.retransmit;
- curr->call_close += ns->tcp.close;
- curr->call_tcp_v4_connection += ns->tcp.ipv4_connect;
- curr->call_tcp_v6_connection += ns->tcp.ipv6_connect;
-
- curr->call_udp_sent += ns->udp.call_udp_sent;
- curr->call_udp_received += ns->udp.call_udp_received;
+ ebpf_socket_publish_apps_t *curr = &local_pid->socket;
+
+ curr->bytes_sent = ns->tcp.tcp_bytes_sent;
+ curr->bytes_received = ns->tcp.tcp_bytes_received;
+ curr->call_tcp_sent = ns->tcp.call_tcp_sent;
+ curr->call_tcp_received = ns->tcp.call_tcp_received;
+ curr->retransmit = ns->tcp.retransmit;
+ curr->call_close = ns->tcp.close;
+ curr->call_tcp_v4_connection = ns->tcp.ipv4_connect;
+ curr->call_tcp_v6_connection = ns->tcp.ipv6_connect;
+
+ curr->call_udp_sent = ns->udp.call_udp_sent;
+ curr->call_udp_received = ns->udp.call_udp_received;
}
/**
@@ -2026,8 +2033,9 @@ static void ebpf_update_socket_cgroup()
for (pids = ect->pids; pids; pids = pids->next) {
int pid = pids->pid;
ebpf_socket_publish_apps_t *publish = &ect->publish_socket;
- if (likely(socket_bandwidth_curr) && socket_bandwidth_curr[pid]) {
- ebpf_socket_publish_apps_t *in = socket_bandwidth_curr[pid];
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ ebpf_socket_publish_apps_t *in = &local_pid->socket;
publish->bytes_sent = in->bytes_sent;
publish->bytes_received = in->bytes_received;
@@ -2329,96 +2337,166 @@ static void ebpf_send_specific_socket_data(char *type, ebpf_socket_publish_apps_
**/
static void ebpf_create_systemd_socket_charts(int update_every)
{
- int order = 20080;
- ebpf_create_charts_on_systemd(NETDATA_NET_APPS_CONNECTION_TCP_V4,
- "Calls to tcp_v4_connection", EBPF_COMMON_DIMENSION_CONNECTIONS,
- NETDATA_APPS_NET_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- order++,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SERVICES_SOCKET_TCP_V4_CONN_CONTEXT, NETDATA_EBPF_MODULE_NAME_SOCKET,
- update_every);
+ static ebpf_systemd_args_t data_tcp_v4 = {
+ .title = "Calls to tcp_v4_connection",
+ .units = EBPF_COMMON_DIMENSION_CONNECTIONS,
+ .family = NETDATA_APPS_NET_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20080,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SERVICES_SOCKET_TCP_V4_CONN_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SOCKET,
+ .update_every = 0,
+ .suffix = NETDATA_NET_APPS_CONNECTION_TCP_V4,
+ .dimension = "connections"
+ };
- if (tcp_v6_connect_address.type == 'T') {
- ebpf_create_charts_on_systemd(NETDATA_NET_APPS_CONNECTION_TCP_V6,
- "Calls to tcp_v6_connection",
- EBPF_COMMON_DIMENSION_CONNECTIONS,
- NETDATA_APPS_NET_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- order++,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SERVICES_SOCKET_TCP_V6_CONN_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_SOCKET,
- update_every);
- }
+ static ebpf_systemd_args_t data_tcp_v6 = {
+ .title = "Calls to tcp_v6_connection",
+ .units = EBPF_COMMON_DIMENSION_CONNECTIONS,
+ .family = NETDATA_APPS_NET_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20081,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SERVICES_SOCKET_TCP_V6_CONN_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SOCKET,
+ .update_every = 0,
+ .suffix = NETDATA_NET_APPS_CONNECTION_TCP_V6,
+ .dimension = "connections"
+ };
- ebpf_create_charts_on_systemd(NETDATA_NET_APPS_BANDWIDTH_RECV,
- "Bytes received", EBPF_COMMON_DIMENSION_BITS,
- NETDATA_APPS_NET_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- order++,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SERVICES_SOCKET_BYTES_RECV_CONTEXT, NETDATA_EBPF_MODULE_NAME_SOCKET,
- update_every);
+ static ebpf_systemd_args_t data_bandwith_recv = {
+ .title = "Bytes received",
+ .units = EBPF_COMMON_DIMENSION_BITS,
+ .family = NETDATA_APPS_NET_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20082,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SERVICES_SOCKET_BYTES_RECV_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SOCKET,
+ .update_every = 0,
+ .suffix = NETDATA_NET_APPS_BANDWIDTH_RECV,
+ .dimension = "bits"
+ };
- ebpf_create_charts_on_systemd(NETDATA_NET_APPS_BANDWIDTH_SENT,
- "Bytes sent", EBPF_COMMON_DIMENSION_BITS,
- NETDATA_APPS_NET_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- order++,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SERVICES_SOCKET_BYTES_SEND_CONTEXT, NETDATA_EBPF_MODULE_NAME_SOCKET,
- update_every);
+ static ebpf_systemd_args_t data_bandwith_sent = {
+ .title = "Bytes sent",
+ .units = EBPF_COMMON_DIMENSION_BITS,
+ .family = NETDATA_APPS_NET_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20083,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SERVICES_SOCKET_BYTES_SEND_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SOCKET,
+ .update_every = 0,
+ .suffix = NETDATA_NET_APPS_BANDWIDTH_SENT,
+ .dimension = "bits"
+ };
- ebpf_create_charts_on_systemd(NETDATA_NET_APPS_BANDWIDTH_TCP_RECV_CALLS,
- "Calls to tcp_cleanup_rbuf.",
- EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_NET_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- order++,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SERVICES_SOCKET_TCP_RECV_CONTEXT, NETDATA_EBPF_MODULE_NAME_SOCKET,
- update_every);
+ static ebpf_systemd_args_t data_tcp_cleanup = {
+ .title = "Calls to tcp_cleanup_rbuf.",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_NET_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20084,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SERVICES_SOCKET_TCP_RECV_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SOCKET,
+ .update_every = 0,
+ .suffix = NETDATA_NET_APPS_BANDWIDTH_TCP_RECV_CALLS,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_NET_APPS_BANDWIDTH_TCP_SEND_CALLS,
- "Calls to tcp_sendmsg.",
- EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_NET_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- order++,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SERVICES_SOCKET_TCP_SEND_CONTEXT, NETDATA_EBPF_MODULE_NAME_SOCKET,
- update_every);
+ static ebpf_systemd_args_t data_tcp_sendmsg = {
+ .title = "Calls to tcp_sendmsg.",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_NET_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20085,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SERVICES_SOCKET_TCP_SEND_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SOCKET,
+ .update_every = 0,
+ .suffix = NETDATA_NET_APPS_BANDWIDTH_TCP_SEND_CALLS,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_NET_APPS_BANDWIDTH_TCP_RETRANSMIT,
- "Calls to tcp_retransmit",
- EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_NET_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- order++,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SERVICES_SOCKET_TCP_RETRANSMIT_CONTEXT, NETDATA_EBPF_MODULE_NAME_SOCKET,
- update_every);
+ static ebpf_systemd_args_t data_tcp_retransmit = {
+ .title = "Calls to tcp_retransmit",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_NET_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20086,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SERVICES_SOCKET_TCP_RETRANSMIT_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SOCKET,
+ .update_every = 0,
+ .suffix = NETDATA_NET_APPS_BANDWIDTH_TCP_RETRANSMIT,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_NET_APPS_BANDWIDTH_UDP_SEND_CALLS,
- "Calls to udp_sendmsg",
- EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_NET_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- order++,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SERVICES_SOCKET_UDP_SEND_CONTEXT, NETDATA_EBPF_MODULE_NAME_SOCKET,
- update_every);
+ static ebpf_systemd_args_t data_udp_send = {
+ .title = "Calls to udp_sendmsg",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_NET_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20087,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SERVICES_SOCKET_UDP_SEND_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SOCKET,
+ .update_every = 0,
+ .suffix = NETDATA_NET_APPS_BANDWIDTH_UDP_SEND_CALLS,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_NET_APPS_BANDWIDTH_UDP_RECV_CALLS,
- "Calls to udp_recvmsg",
- EBPF_COMMON_DIMENSION_CALL,
- NETDATA_APPS_NET_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED,
- order++,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SERVICES_SOCKET_UDP_RECV_CONTEXT, NETDATA_EBPF_MODULE_NAME_SOCKET,
- update_every);
+ static ebpf_systemd_args_t data_udp_recv = {
+ .title = "Calls to udp_recvmsg",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_APPS_NET_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20088,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SERVICES_SOCKET_UDP_RECV_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SOCKET,
+ .update_every = 0,
+ .suffix = NETDATA_NET_APPS_BANDWIDTH_UDP_RECV_CALLS,
+ .dimension = "calls"
+ };
+
+ if (!data_tcp_v4.update_every)
+ data_tcp_v4.update_every = data_tcp_v6.update_every = data_bandwith_recv.update_every =
+ data_bandwith_sent.update_every = data_tcp_cleanup.update_every = data_tcp_sendmsg.update_every =
+ data_tcp_retransmit.update_every = data_udp_send.update_every = data_udp_recv.update_every = update_every;
+
+ ebpf_cgroup_target_t *w;
+ for (w = ebpf_cgroup_pids; w ; w = w->next) {
+ if (unlikely(!w->systemd || w->flags & NETDATA_EBPF_SERVICES_HAS_SOCKET_CHART))
+ continue;
+
+ data_tcp_v4.id = data_tcp_v6.id = data_bandwith_recv.id =
+ data_bandwith_sent.id = data_tcp_cleanup.id = data_tcp_sendmsg.id =
+ data_tcp_retransmit.id = data_udp_send.id = data_udp_recv.id = w->name;
+
+ ebpf_create_charts_on_systemd(&data_tcp_v4);
+ if (tcp_v6_connect_address.type == 'T') {
+ ebpf_create_charts_on_systemd(&data_tcp_v6);
+ }
+
+ ebpf_create_charts_on_systemd(&data_bandwith_recv);
+ ebpf_create_charts_on_systemd(&data_bandwith_sent);
+
+ ebpf_create_charts_on_systemd(&data_tcp_cleanup);
+
+ ebpf_create_charts_on_systemd(&data_tcp_sendmsg);
+
+ ebpf_create_charts_on_systemd(&data_tcp_retransmit);
+
+ ebpf_create_charts_on_systemd(&data_udp_recv);
+
+ ebpf_create_charts_on_systemd(&data_udp_send);
+
+ w->flags |= NETDATA_EBPF_SERVICES_HAS_SOCKET_CHART;
+ }
}
/**
@@ -2429,79 +2507,49 @@ static void ebpf_create_systemd_socket_charts(int update_every)
static void ebpf_send_systemd_socket_charts()
{
ebpf_cgroup_target_t *ect;
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_NET_APPS_CONNECTION_TCP_V4, "");
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_socket.call_tcp_v4_connection);
+ if (unlikely(!(ect->flags & NETDATA_EBPF_SERVICES_HAS_SOCKET_CHART)) ) {
+ continue;
}
- }
- ebpf_write_end_chart();
- if (tcp_v6_connect_address.type == 'T') {
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_NET_APPS_CONNECTION_TCP_V6, "");
- for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_socket.call_tcp_v6_connection);
- }
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_NET_APPS_CONNECTION_TCP_V4);
+ write_chart_dimension("connections", (long long)ect->publish_socket.call_tcp_v4_connection);
ebpf_write_end_chart();
- }
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_NET_APPS_BANDWIDTH_SENT, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_socket.bytes_sent);
+ if (tcp_v6_connect_address.type == 'T') {
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_NET_APPS_CONNECTION_TCP_V6);
+ write_chart_dimension("connections", (long long)ect->publish_socket.call_tcp_v6_connection);
+ ebpf_write_end_chart();
}
- }
- ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_NET_APPS_BANDWIDTH_RECV, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_socket.bytes_received);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_NET_APPS_BANDWIDTH_SENT);
+ write_chart_dimension("bits", (long long)ect->publish_socket.bytes_sent);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_NET_APPS_BANDWIDTH_TCP_SEND_CALLS, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_socket.call_tcp_sent);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_NET_APPS_BANDWIDTH_RECV);
+ write_chart_dimension("bits", (long long)ect->publish_socket.bytes_received);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_NET_APPS_BANDWIDTH_TCP_RECV_CALLS, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_socket.call_tcp_received);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_NET_APPS_BANDWIDTH_TCP_SEND_CALLS);
+ write_chart_dimension("calls", (long long)ect->publish_socket.call_tcp_sent);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_NET_APPS_BANDWIDTH_TCP_RETRANSMIT, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_socket.retransmit);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_NET_APPS_BANDWIDTH_TCP_RECV_CALLS);
+ write_chart_dimension("calls", (long long)ect->publish_socket.call_tcp_received);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_NET_APPS_BANDWIDTH_UDP_SEND_CALLS, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_socket.call_udp_sent);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_NET_APPS_BANDWIDTH_TCP_RETRANSMIT);
+ write_chart_dimension("calls", (long long)ect->publish_socket.retransmit);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_NET_APPS_BANDWIDTH_UDP_RECV_CALLS, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long)ect->publish_socket.call_udp_received);
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_NET_APPS_BANDWIDTH_UDP_SEND_CALLS);
+ write_chart_dimension("calls", (long long)ect->publish_socket.call_udp_sent);
+ ebpf_write_end_chart();
+
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_NET_APPS_BANDWIDTH_UDP_RECV_CALLS);
+ write_chart_dimension("calls", (long long)ect->publish_socket.call_udp_received);
+ ebpf_write_end_chart();
}
- ebpf_write_end_chart();
}
/**
@@ -2525,17 +2573,13 @@ void ebpf_socket_update_cgroup_algorithm()
*/
static void ebpf_socket_send_cgroup_data(int update_every)
{
- if (!ebpf_cgroup_pids)
- return;
-
pthread_mutex_lock(&mutex_cgroup_shm);
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
ebpf_socket_sum_cgroup_pids(&ect->publish_socket, ect->pids);
}
- int has_systemd = shm_ebpf_cgroup.header->systemd_enabled;
- if (has_systemd) {
+ if (shm_ebpf_cgroup.header->systemd_enabled) {
if (send_cgroup_chart) {
ebpf_create_systemd_socket_charts(update_every);
}
@@ -2602,7 +2646,6 @@ static void socket_collector(ebpf_module_t *em)
ebpf_socket_read_hash_global_tables(stats, maps_per_core);
}
- pthread_mutex_lock(&collect_data_mutex);
if (cgroups)
ebpf_update_socket_cgroup();
@@ -2611,20 +2654,14 @@ static void socket_collector(ebpf_module_t *em)
ebpf_socket_send_data(em);
if (socket_apps_enabled & NETDATA_EBPF_APPS_FLAG_CHART_CREATED)
- ebpf_socket_send_apps_data(em, apps_groups_root_target);
-
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_socket_pid)
- ebpf_send_data_aral_chart(ebpf_aral_socket_pid, em);
-#endif
+ ebpf_socket_send_apps_data();
- if (cgroups)
+ if (cgroups && shm_ebpf_cgroup.header && ebpf_cgroup_pids)
ebpf_socket_send_cgroup_data(update_every);
fflush(stdout);
pthread_mutex_unlock(&lock);
- pthread_mutex_unlock(&collect_data_mutex);
pthread_mutex_lock(&ebpf_exit_cleanup);
if (running_time && !em->running_time)
@@ -2655,9 +2692,6 @@ static void ebpf_socket_initialize_global_vectors()
memset(socket_publish_aggregated, 0 ,NETDATA_MAX_SOCKET_VECTOR * sizeof(netdata_publish_syscall_t));
socket_hash_values = callocz(ebpf_nprocs, sizeof(netdata_idx_t));
- ebpf_socket_aral_init();
- socket_bandwidth_curr = callocz((size_t)pid_max, sizeof(ebpf_socket_publish_apps_t *));
-
aral_socket_table = ebpf_allocate_pid_aral(NETDATA_EBPF_SOCKET_ARAL_TABLE_NAME,
sizeof(netdata_socket_plus_t));
@@ -2878,11 +2912,6 @@ void *ebpf_socket_thread(void *ptr)
ebpf_update_stats(&plugin_statistics, em);
ebpf_update_kernel_memory_with_vector(&plugin_statistics, em->maps, EBPF_ACTION_STAT_ADD);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_socket_pid)
- socket_disable_priority = ebpf_statistic_create_aral_chart(NETDATA_EBPF_SOCKET_ARAL_NAME, em);
-#endif
-
pthread_mutex_unlock(&lock);
socket_collector(em);
diff --git a/collectors/ebpf.plugin/ebpf_socket.h b/src/collectors/ebpf.plugin/ebpf_socket.h
index a6d3e03b6..64201149f 100644
--- a/collectors/ebpf.plugin/ebpf_socket.h
+++ b/src/collectors/ebpf.plugin/ebpf_socket.h
@@ -95,8 +95,7 @@ typedef enum ebpf_socket_idx {
} ebpf_socket_index_t;
#define NETDATA_SOCKET_KERNEL_FUNCTIONS "kernel"
-#define NETDATA_NETWORK_CONNECTIONS_GROUP "network connections"
-#define NETDATA_CGROUP_NET_GROUP "network (eBPF)"
+#define NETDATA_CGROUP_NET_GROUP "network"
// Global chart name
#define NETDATA_TCP_OUTBOUND_CONNECTIONS "tcp_outbound_conn"
@@ -142,15 +141,15 @@ typedef enum ebpf_socket_idx {
#define NETDATA_CGROUP_SOCKET_UDP_RECV_CONTEXT "cgroup.net_udp_recv"
#define NETDATA_CGROUP_SOCKET_UDP_SEND_CONTEXT "cgroup.net_udp_send"
-#define NETDATA_SERVICES_SOCKET_TCP_V4_CONN_CONTEXT "services.net_conn_ipv4"
-#define NETDATA_SERVICES_SOCKET_TCP_V6_CONN_CONTEXT "services.net_conn_ipv6"
-#define NETDATA_SERVICES_SOCKET_BYTES_RECV_CONTEXT "services.net_bytes_recv"
-#define NETDATA_SERVICES_SOCKET_BYTES_SEND_CONTEXT "services.net_bytes_send"
-#define NETDATA_SERVICES_SOCKET_TCP_RECV_CONTEXT "services.net_tcp_recv"
-#define NETDATA_SERVICES_SOCKET_TCP_SEND_CONTEXT "services.net_tcp_send"
-#define NETDATA_SERVICES_SOCKET_TCP_RETRANSMIT_CONTEXT "services.net_retransmit"
-#define NETDATA_SERVICES_SOCKET_UDP_RECV_CONTEXT "services.net_udp_recv"
-#define NETDATA_SERVICES_SOCKET_UDP_SEND_CONTEXT "services.net_udp_send"
+#define NETDATA_SERVICES_SOCKET_TCP_V4_CONN_CONTEXT "systemd.services.net_conn_ipv4"
+#define NETDATA_SERVICES_SOCKET_TCP_V6_CONN_CONTEXT "systemd.services.net_conn_ipv6"
+#define NETDATA_SERVICES_SOCKET_BYTES_RECV_CONTEXT "systemd.services.net_bytes_recv"
+#define NETDATA_SERVICES_SOCKET_BYTES_SEND_CONTEXT "systemd.services.net_bytes_send"
+#define NETDATA_SERVICES_SOCKET_TCP_RECV_CONTEXT "systemd.services.net_tcp_recv"
+#define NETDATA_SERVICES_SOCKET_TCP_SEND_CONTEXT "systemd.services.net_tcp_send"
+#define NETDATA_SERVICES_SOCKET_TCP_RETRANSMIT_CONTEXT "systemd.services.net_retransmit"
+#define NETDATA_SERVICES_SOCKET_UDP_RECV_CONTEXT "systemd.services.net_udp_recv"
+#define NETDATA_SERVICES_SOCKET_UDP_SEND_CONTEXT "systemd.services.net_udp_send"
// ARAL name
#define NETDATA_EBPF_SOCKET_ARAL_NAME "ebpf_socket"
@@ -269,6 +268,8 @@ extern ebpf_network_viewer_options_t network_viewer_opt;
* Structure to store socket information
*/
typedef struct netdata_socket {
+ char name[TASK_COMM_LEN];
+
// Timestamp
uint64_t first_timestamp;
uint64_t current_timestamp;
@@ -285,6 +286,7 @@ typedef struct netdata_socket {
uint32_t retransmit; //It is never used with UDP
uint32_t ipv4_connect;
uint32_t ipv6_connect;
+ uint32_t state; // We do not have charts for it, because we are using network viewer plugin
} tcp;
struct {
diff --git a/collectors/ebpf.plugin/ebpf_softirq.c b/src/collectors/ebpf.plugin/ebpf_softirq.c
index 106ff4f29..106ff4f29 100644
--- a/collectors/ebpf.plugin/ebpf_softirq.c
+++ b/src/collectors/ebpf.plugin/ebpf_softirq.c
diff --git a/collectors/ebpf.plugin/ebpf_softirq.h b/src/collectors/ebpf.plugin/ebpf_softirq.h
index 4ef36775a..4ef36775a 100644
--- a/collectors/ebpf.plugin/ebpf_softirq.h
+++ b/src/collectors/ebpf.plugin/ebpf_softirq.h
diff --git a/collectors/ebpf.plugin/ebpf_swap.c b/src/collectors/ebpf.plugin/ebpf_swap.c
index fb007f928..42aa04ad8 100644
--- a/collectors/ebpf.plugin/ebpf_swap.c
+++ b/src/collectors/ebpf.plugin/ebpf_swap.c
@@ -52,6 +52,17 @@ netdata_ebpf_targets_t swap_targets[] = { {.name = "swap_readpage", .mode = EBPF
{.name = "swap_writepage", .mode = EBPF_LOAD_TRAMPOLINE},
{.name = NULL, .mode = EBPF_LOAD_TRAMPOLINE}};
+struct netdata_static_thread ebpf_read_swap = {
+ .name = "EBPF_READ_SWAP",
+ .config_section = NULL,
+ .config_name = NULL,
+ .env_name = NULL,
+ .enabled = 1,
+ .thread = NULL,
+ .init_routine = NULL,
+ .start_routine = NULL
+};
+
#ifdef LIBBPF_MAJOR_VERSION
/**
* Disable probe
@@ -77,7 +88,6 @@ static void ebpf_swap_disable_trampoline(struct swap_bpf *obj)
{
bpf_program__set_autoload(obj->progs.netdata_swap_readpage_fentry, false);
bpf_program__set_autoload(obj->progs.netdata_swap_writepage_fentry, false);
- bpf_program__set_autoload(obj->progs.netdata_release_task_fentry, false);
}
/**
@@ -94,9 +104,6 @@ static void ebpf_swap_set_trampoline_target(struct swap_bpf *obj)
bpf_program__set_attach_target(obj->progs.netdata_swap_writepage_fentry, 0,
swap_targets[NETDATA_KEY_SWAP_WRITEPAGE_CALL].name);
-
- bpf_program__set_attach_target(obj->progs.netdata_release_task_fentry, 0,
- EBPF_COMMON_FNCT_CLEAN_UP);
}
/**
@@ -160,18 +167,6 @@ static void ebpf_swap_adjust_map(struct swap_bpf *obj, ebpf_module_t *em)
}
/**
- * Disable Release Task
- *
- * Disable release task when apps is not enabled.
- *
- * @param obj is the main structure for bpf objects.
- */
-static void ebpf_swap_disable_release_task(struct swap_bpf *obj)
-{
- bpf_program__set_autoload(obj->progs.netdata_release_task_fentry, false);
-}
-
-/**
* Load and attach
*
* Load and attach the eBPF code in kernel.
@@ -196,9 +191,6 @@ static inline int ebpf_swap_load_and_attach(struct swap_bpf *obj, ebpf_module_t
ebpf_swap_adjust_map(obj, em);
- if (!em->apps_charts && !em->cgroup_charts)
- ebpf_swap_disable_release_task(obj);
-
int ret = swap_bpf__load(obj);
if (ret) {
return ret;
@@ -230,25 +222,25 @@ static void ebpf_obsolete_specific_swap_charts(char *type, int update_every);
*
* @param em a pointer to `struct ebpf_module`
*/
-static void ebpf_obsolete_swap_services(ebpf_module_t *em)
+static void ebpf_obsolete_swap_services(ebpf_module_t *em, char *id)
{
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_MEM_SWAP_READ_CHART,
- "",
"Calls to function swap_readpage.",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_SYSTEM_CGROUP_SWAP_SUBMENU,
+ NETDATA_SYSTEM_SWAP_SUBMENU,
NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_CGROUP_SWAP_READ_CONTEXT,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5100,
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_MEM_SWAP_WRITE_CHART,
- "",
"Calls to function swap_writepage.",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_SYSTEM_CGROUP_SWAP_SUBMENU,
+ NETDATA_SYSTEM_SWAP_SUBMENU,
NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_CGROUP_SWAP_WRITE_CONTEXT,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5101,
@@ -265,12 +257,13 @@ static void ebpf_obsolete_swap_services(ebpf_module_t *em)
static inline void ebpf_obsolete_swap_cgroup_charts(ebpf_module_t *em) {
pthread_mutex_lock(&mutex_cgroup_shm);
- ebpf_obsolete_swap_services(em);
-
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (ect->systemd)
+ if (ect->systemd) {
+ ebpf_obsolete_swap_services(em, ect->name);
+
continue;
+ }
ebpf_obsolete_specific_swap_charts(ect->name, em->update_every);
}
@@ -288,6 +281,7 @@ void ebpf_obsolete_swap_apps_charts(struct ebpf_module *em)
{
struct ebpf_target *w;
int update_every = em->update_every;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = apps_groups_root_target; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_SWAP_IDX))))
continue;
@@ -315,6 +309,7 @@ void ebpf_obsolete_swap_apps_charts(struct ebpf_module *em)
update_every);
w->charts_created &= ~(1<<EBPF_MODULE_SWAP_IDX);
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -348,6 +343,9 @@ static void ebpf_swap_exit(void *ptr)
{
ebpf_module_t *em = (ebpf_module_t *)ptr;
+ if (ebpf_read_swap.thread)
+ netdata_thread_cancel(*ebpf_read_swap.thread);
+
if (em->enabled == NETDATA_THREAD_EBPF_FUNCTION_RUNNING) {
pthread_mutex_lock(&lock);
if (em->cgroup_charts) {
@@ -411,59 +409,24 @@ static void swap_apps_accumulator(netdata_publish_swap_t *out, int maps_per_core
}
/**
- * Fill PID
- *
- * Fill PID structures
- *
- * @param current_pid pid that we are collecting data
- * @param out values read from hash tables;
- */
-static void swap_fill_pid(uint32_t current_pid, netdata_publish_swap_t *publish)
-{
- netdata_publish_swap_t *curr = swap_pid[current_pid];
- if (!curr) {
- curr = callocz(1, sizeof(netdata_publish_swap_t));
- swap_pid[current_pid] = curr;
- }
-
- memcpy(curr, publish, sizeof(netdata_publish_swap_t));
-}
-
-/**
* Update cgroup
*
* Update cgroup data based in
- *
- * @param maps_per_core do I need to read all cores?
*/
-static void ebpf_update_swap_cgroup(int maps_per_core)
+static void ebpf_update_swap_cgroup()
{
ebpf_cgroup_target_t *ect ;
- netdata_publish_swap_t *cv = swap_vector;
- int fd = swap_maps[NETDATA_PID_SWAP_TABLE].map_fd;
- size_t length = sizeof(netdata_publish_swap_t);
- if (maps_per_core)
- length *= ebpf_nprocs;
pthread_mutex_lock(&mutex_cgroup_shm);
for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
struct pid_on_target2 *pids;
for (pids = ect->pids; pids; pids = pids->next) {
int pid = pids->pid;
netdata_publish_swap_t *out = &pids->swap;
- if (likely(swap_pid) && swap_pid[pid]) {
- netdata_publish_swap_t *in = swap_pid[pid];
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ netdata_publish_swap_t *in = &local_pid->swap;
memcpy(out, in, sizeof(netdata_publish_swap_t));
- } else {
- memset(cv, 0, length);
- if (!bpf_map_lookup_elem(fd, &pid, cv)) {
- swap_apps_accumulator(cv, maps_per_core);
-
- memcpy(out, cv, sizeof(netdata_publish_swap_t));
-
- // We are cleaning to avoid passing data read from one process to other.
- memset(cv, 0, length);
- }
}
}
}
@@ -471,38 +434,143 @@ static void ebpf_update_swap_cgroup(int maps_per_core)
}
/**
+ * Sum PIDs
+ *
+ * Sum values for all targets.
+ *
+ * @param swap
+ * @param root
+ */
+static void ebpf_swap_sum_pids(netdata_publish_swap_t *swap, struct ebpf_pid_on_target *root)
+{
+ uint64_t local_read = 0;
+ uint64_t local_write = 0;
+
+ while (root) {
+ int32_t pid = root->pid;
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ netdata_publish_swap_t *w = &local_pid->swap;
+ local_write += w->write;
+ local_read += w->read;
+ }
+ root = root->next;
+ }
+
+ // These conditions were added, because we are using incremental algorithm
+ swap->write = (local_write >= swap->write) ? local_write : swap->write;
+ swap->read = (local_read >= swap->read) ? local_read : swap->read;
+ }
+
+
+/**
+ * Resume apps data
+ */
+void ebpf_swap_resume_apps_data() {
+ struct ebpf_target *w;
+ for (w = apps_groups_root_target; w; w = w->next) {
+ if (unlikely(!(w->charts_created & (1 << EBPF_MODULE_SWAP_IDX))))
+ continue;
+
+ ebpf_swap_sum_pids(&w->swap, w->root_pid);
+ }
+}
+
+/**
* Read APPS table
*
* Read the apps table and store data inside the structure.
*
* @param maps_per_core do I need to read all cores?
*/
-static void read_swap_apps_table(int maps_per_core)
+static void ebpf_read_swap_apps_table(int maps_per_core, int max_period)
{
netdata_publish_swap_t *cv = swap_vector;
- uint32_t key;
- struct ebpf_pid_stat *pids = ebpf_root_of_pids;
int fd = swap_maps[NETDATA_PID_SWAP_TABLE].map_fd;
size_t length = sizeof(netdata_publish_swap_t);
if (maps_per_core)
length *= ebpf_nprocs;
- while (pids) {
- key = pids->pid;
+ uint32_t key = 0, next_key = 0;
+ while (bpf_map_get_next_key(fd, &key, &next_key) == 0) {
if (bpf_map_lookup_elem(fd, &key, cv)) {
- pids = pids->next;
- continue;
+ goto end_swap_loop;
}
swap_apps_accumulator(cv, maps_per_core);
- swap_fill_pid(key, cv);
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(key, cv->tgid);
+ if (!local_pid)
+ goto end_swap_loop;
+
+ netdata_publish_swap_t *publish = &local_pid->swap;
+ if (!publish->ct || publish->ct != cv->ct) {
+ memcpy(publish, cv, sizeof(netdata_publish_swap_t));
+ local_pid->not_updated = 0;
+ } else if (++local_pid->not_updated >= max_period) {
+ bpf_map_delete_elem(fd, &key);
+ local_pid->not_updated = 0;
+ }
// We are cleaning to avoid passing data read from one process to other.
+end_swap_loop:
memset(cv, 0, length);
+ key = next_key;
+ }
+}
- pids = pids->next;
+/**
+ * SWAP thread
+ *
+ * Thread used to generate swap charts.
+ *
+ * @param ptr a pointer to `struct ebpf_module`
+ *
+ * @return It always return NULL
+ */
+void *ebpf_read_swap_thread(void *ptr)
+{
+ heartbeat_t hb;
+ heartbeat_init(&hb);
+
+ ebpf_module_t *em = (ebpf_module_t *)ptr;
+
+ int maps_per_core = em->maps_per_core;
+ int update_every = em->update_every;
+
+ int counter = update_every - 1;
+
+ uint32_t lifetime = em->lifetime;
+ uint32_t running_time = 0;
+ usec_t period = update_every * USEC_PER_SEC;
+ int max_period = update_every * EBPF_CLEANUP_FACTOR;
+
+ while (!ebpf_plugin_exit && running_time < lifetime) {
+ (void)heartbeat_next(&hb, period);
+ if (ebpf_plugin_exit || ++counter != update_every)
+ continue;
+
+ netdata_thread_disable_cancelability();
+
+ pthread_mutex_lock(&collect_data_mutex);
+ ebpf_read_swap_apps_table(maps_per_core, max_period);
+ ebpf_swap_resume_apps_data();
+ pthread_mutex_unlock(&collect_data_mutex);
+
+ counter = 0;
+
+ pthread_mutex_lock(&ebpf_exit_cleanup);
+ if (running_time && !em->running_time)
+ running_time = update_every;
+ else
+ running_time += update_every;
+
+ em->running_time = running_time;
+ pthread_mutex_unlock(&ebpf_exit_cleanup);
+ netdata_thread_enable_cancelability();
}
+
+ return NULL;
}
/**
@@ -545,34 +613,6 @@ static void ebpf_swap_read_global_table(netdata_idx_t *stats, int maps_per_core)
}
/**
- * Sum PIDs
- *
- * Sum values for all targets.
- *
- * @param swap
- * @param root
- */
-static void ebpf_swap_sum_pids(netdata_publish_swap_t *swap, struct ebpf_pid_on_target *root)
-{
- uint64_t local_read = 0;
- uint64_t local_write = 0;
-
- while (root) {
- int32_t pid = root->pid;
- netdata_publish_swap_t *w = swap_pid[pid];
- if (w) {
- local_write += w->write;
- local_read += w->read;
- }
- root = root->next;
- }
-
- // These conditions were added, because we are using incremental algorithm
- swap->write = (local_write >= swap->write) ? local_write : swap->write;
- swap->read = (local_read >= swap->read) ? local_read : swap->read;
-}
-
-/**
* Send data to Netdata calling auxiliary functions.
*
* @param root the target list.
@@ -580,12 +620,11 @@ static void ebpf_swap_sum_pids(netdata_publish_swap_t *swap, struct ebpf_pid_on_
void ebpf_swap_send_apps_data(struct ebpf_target *root)
{
struct ebpf_target *w;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = root; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_SWAP_IDX))))
continue;
- ebpf_swap_sum_pids(&w->swap, w->root_pid);
-
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_call_swap_readpage");
write_chart_dimension("calls", (long long) w->swap.read);
ebpf_write_end_chart();
@@ -594,6 +633,7 @@ void ebpf_swap_send_apps_data(struct ebpf_target *root)
write_chart_dimension("calls", (long long) w->swap.write);
ebpf_write_end_chart();
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -630,21 +670,19 @@ static void ebpf_swap_sum_cgroup_pids(netdata_publish_swap_t *swap, struct pid_o
static void ebpf_send_systemd_swap_charts()
{
ebpf_cgroup_target_t *ect;
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_MEM_SWAP_READ_CHART, "");
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long) ect->publish_systemd_swap.read);
+ if (unlikely(!(ect->flags & NETDATA_EBPF_SERVICES_HAS_SWAP_CHART)) ) {
+ continue;
}
- }
- ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_MEM_SWAP_WRITE_CHART, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, (long long) ect->publish_systemd_swap.write);
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_MEM_SWAP_READ_CHART);
+ write_chart_dimension("calls", (long long) ect->publish_systemd_swap.read);
+ ebpf_write_end_chart();
+
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_MEM_SWAP_WRITE_CHART);
+ write_chart_dimension("calls", (long long) ect->publish_systemd_swap.write);
+ ebpf_write_end_chart();
}
- ebpf_write_end_chart();
}
/**
@@ -659,7 +697,7 @@ static void ebpf_create_specific_swap_charts(char *type, int update_every)
{
ebpf_create_chart(type, NETDATA_MEM_SWAP_READ_CHART,
"Calls to function swap_readpage.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_SYSTEM_CGROUP_SWAP_SUBMENU,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_SYSTEM_SWAP_SUBMENU,
NETDATA_CGROUP_SWAP_READ_CONTEXT, NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5100,
ebpf_create_global_dimension,
@@ -667,7 +705,7 @@ static void ebpf_create_specific_swap_charts(char *type, int update_every)
ebpf_create_chart(type, NETDATA_MEM_SWAP_WRITE_CHART,
"Calls to function swap_writepage.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_SYSTEM_CGROUP_SWAP_SUBMENU,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_SYSTEM_SWAP_SUBMENU,
NETDATA_CGROUP_SWAP_WRITE_CONTEXT, NETDATA_EBPF_CHART_TYPE_LINE,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5101,
ebpf_create_global_dimension,
@@ -686,12 +724,12 @@ static void ebpf_create_specific_swap_charts(char *type, int update_every)
static void ebpf_obsolete_specific_swap_charts(char *type, int update_every)
{
ebpf_write_chart_obsolete(type, NETDATA_MEM_SWAP_READ_CHART, "", "Calls to function swap_readpage.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_SYSTEM_CGROUP_SWAP_SUBMENU,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_SYSTEM_SWAP_SUBMENU,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CGROUP_SWAP_READ_CONTEXT,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5100, update_every);
ebpf_write_chart_obsolete(type, NETDATA_MEM_SWAP_WRITE_CHART, "", "Calls to function swap_writepage.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_SYSTEM_CGROUP_SWAP_SUBMENU,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_SYSTEM_SWAP_SUBMENU,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CGROUP_SWAP_WRITE_CONTEXT,
NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5101, update_every);
}
@@ -724,19 +762,49 @@ static void ebpf_send_specific_swap_data(char *type, netdata_publish_swap_t *val
**/
static void ebpf_create_systemd_swap_charts(int update_every)
{
- ebpf_create_charts_on_systemd(NETDATA_MEM_SWAP_READ_CHART,
- "Calls to swap_readpage.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_SYSTEM_CGROUP_SWAP_SUBMENU,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20191,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_SWAP_READ_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_SWAP, update_every);
-
- ebpf_create_charts_on_systemd(NETDATA_MEM_SWAP_WRITE_CHART,
- "Calls to function swap_writepage.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_SYSTEM_CGROUP_SWAP_SUBMENU,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20192,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_SWAP_WRITE_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_SWAP, update_every);
+ static ebpf_systemd_args_t data_read = {
+ .title = "Calls to swap_readpage.",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_SYSTEM_SWAP_SUBMENU,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20191,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_SWAP_READ_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SWAP,
+ .update_every = 0,
+ .suffix = NETDATA_MEM_SWAP_READ_CHART,
+ .dimension = "calls"
+ };
+
+ static ebpf_systemd_args_t data_write = {
+ .title = "Calls to function swap_writepage.",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_SYSTEM_SWAP_SUBMENU,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20192,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_SWAP_WRITE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_SWAP,
+ .update_every = 0,
+ .suffix = NETDATA_MEM_SWAP_WRITE_CHART,
+ .dimension = "calls"
+ };
+
+ if (!data_write.update_every)
+ data_read.update_every = data_write.update_every = update_every;
+
+ ebpf_cgroup_target_t *w;
+ for (w = ebpf_cgroup_pids; w ; w = w->next) {
+ if (unlikely(!w->systemd || w->flags & NETDATA_EBPF_SERVICES_HAS_SWAP_CHART))
+ continue;
+
+ data_read.id = data_write.id = w->name;
+ ebpf_create_charts_on_systemd(&data_read);
+
+ ebpf_create_charts_on_systemd(&data_write);
+
+ w->flags |= NETDATA_EBPF_SERVICES_HAS_SWAP_CHART;
+ }
}
/**
@@ -746,18 +814,13 @@ static void ebpf_create_systemd_swap_charts(int update_every)
*/
void ebpf_swap_send_cgroup_data(int update_every)
{
- if (!ebpf_cgroup_pids)
- return;
-
pthread_mutex_lock(&mutex_cgroup_shm);
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
ebpf_swap_sum_cgroup_pids(&ect->publish_systemd_swap, ect->pids);
}
- int has_systemd = shm_ebpf_cgroup.header->systemd_enabled;
-
- if (has_systemd) {
+ if (shm_ebpf_cgroup.header->systemd_enabled) {
if (send_cgroup_chart) {
ebpf_create_systemd_swap_charts(update_every);
fflush(stdout);
@@ -810,12 +873,9 @@ static void swap_collector(ebpf_module_t *em)
counter = 0;
netdata_apps_integration_flags_t apps = em->apps_charts;
ebpf_swap_read_global_table(stats, maps_per_core);
- pthread_mutex_lock(&collect_data_mutex);
- if (apps)
- read_swap_apps_table(maps_per_core);
if (cgroup)
- ebpf_update_swap_cgroup(maps_per_core);
+ ebpf_update_swap_cgroup();
pthread_mutex_lock(&lock);
@@ -824,11 +884,10 @@ static void swap_collector(ebpf_module_t *em)
if (apps & NETDATA_EBPF_APPS_FLAG_CHART_CREATED)
ebpf_swap_send_apps_data(apps_groups_root_target);
- if (cgroup)
+ if (cgroup && shm_ebpf_cgroup.header && ebpf_cgroup_pids)
ebpf_swap_send_cgroup_data(update_every);
pthread_mutex_unlock(&lock);
- pthread_mutex_unlock(&collect_data_mutex);
pthread_mutex_lock(&ebpf_exit_cleanup);
if (running_time && !em->running_time)
@@ -903,14 +962,9 @@ void ebpf_swap_create_apps_charts(struct ebpf_module *em, void *ptr)
*
* We are not testing the return, because callocz does this and shutdown the software
* case it was not possible to allocate.
- *
- * @param apps is apps enabled?
*/
-static void ebpf_swap_allocate_global_vectors(int apps)
+static void ebpf_swap_allocate_global_vectors()
{
- if (apps)
- swap_pid = callocz((size_t)pid_max, sizeof(netdata_publish_swap_t *));
-
swap_vector = callocz((size_t)ebpf_nprocs, sizeof(netdata_publish_swap_t));
swap_values = callocz((size_t)ebpf_nprocs, sizeof(netdata_idx_t));
@@ -1008,7 +1062,7 @@ void *ebpf_swap_thread(void *ptr)
goto endswap;
}
- ebpf_swap_allocate_global_vectors(em->apps_charts);
+ ebpf_swap_allocate_global_vectors();
int algorithms[NETDATA_SWAP_END] = { NETDATA_EBPF_INCREMENTAL_IDX, NETDATA_EBPF_INCREMENTAL_IDX };
ebpf_global_labels(swap_aggregated_data, swap_publish_aggregated, swap_dimension_name, swap_dimension_name,
@@ -1020,6 +1074,13 @@ void *ebpf_swap_thread(void *ptr)
ebpf_update_kernel_memory_with_vector(&plugin_statistics, em->maps, EBPF_ACTION_STAT_ADD);
pthread_mutex_unlock(&lock);
+ ebpf_read_swap.thread = mallocz(sizeof(netdata_thread_t));
+ netdata_thread_create(ebpf_read_swap.thread,
+ ebpf_read_swap.name,
+ NETDATA_THREAD_OPTION_DEFAULT,
+ ebpf_read_swap_thread,
+ em);
+
swap_collector(em);
endswap:
diff --git a/collectors/ebpf.plugin/ebpf_swap.h b/src/collectors/ebpf.plugin/ebpf_swap.h
index 79e9a01ac..38e700ee0 100644
--- a/collectors/ebpf.plugin/ebpf_swap.h
+++ b/src/collectors/ebpf.plugin/ebpf_swap.h
@@ -19,12 +19,18 @@
#define NETDATA_DIRECTORY_SWAP_CONFIG_FILE "swap.conf"
// Contexts
-#define NETDATA_CGROUP_SWAP_READ_CONTEXT "cgroup.swap_read"
-#define NETDATA_CGROUP_SWAP_WRITE_CONTEXT "cgroup.swap_write"
-#define NETDATA_SYSTEMD_SWAP_READ_CONTEXT "services.swap_read"
-#define NETDATA_SYSTEMD_SWAP_WRITE_CONTEXT "services.swap_write"
+#define NETDATA_CGROUP_SWAP_READ_CONTEXT "systemd.cgroup.swap_read"
+#define NETDATA_CGROUP_SWAP_WRITE_CONTEXT "systemd.cgroup.swap_write"
+#define NETDATA_SYSTEMD_SWAP_READ_CONTEXT "systemd.services.swap_read"
+#define NETDATA_SYSTEMD_SWAP_WRITE_CONTEXT "systemd.services.swap_write"
typedef struct netdata_publish_swap {
+ uint64_t ct;
+ uint32_t tgid;
+ uint32_t uid;
+ uint32_t gid;
+ char name[TASK_COMM_LEN];
+
uint64_t read;
uint64_t write;
} netdata_publish_swap_t;
diff --git a/collectors/ebpf.plugin/ebpf_sync.c b/src/collectors/ebpf.plugin/ebpf_sync.c
index a16318107..a16318107 100644
--- a/collectors/ebpf.plugin/ebpf_sync.c
+++ b/src/collectors/ebpf.plugin/ebpf_sync.c
diff --git a/collectors/ebpf.plugin/ebpf_sync.h b/src/collectors/ebpf.plugin/ebpf_sync.h
index bd1bb78b0..373695565 100644
--- a/collectors/ebpf.plugin/ebpf_sync.h
+++ b/src/collectors/ebpf.plugin/ebpf_sync.h
@@ -3,10 +3,6 @@
#ifndef NETDATA_EBPF_SYNC_H
#define NETDATA_EBPF_SYNC_H 1
-#ifdef LIBBPF_MAJOR_VERSION
-#include "includes/sync.skel.h"
-#endif
-
// Module name & description
#define NETDATA_EBPF_MODULE_NAME_SYNC "sync"
#define NETDATA_EBPF_SYNC_MODULE_DESC "Monitor calls to syscalls sync(2), fsync(2), fdatasync(2), syncfs(2), msync(2), and sync_file_range(2)."
diff --git a/collectors/ebpf.plugin/ebpf_unittest.c b/src/collectors/ebpf.plugin/ebpf_unittest.c
index 11b449e03..11b449e03 100644
--- a/collectors/ebpf.plugin/ebpf_unittest.c
+++ b/src/collectors/ebpf.plugin/ebpf_unittest.c
diff --git a/collectors/ebpf.plugin/ebpf_unittest.h b/src/collectors/ebpf.plugin/ebpf_unittest.h
index 429cbe628..429cbe628 100644
--- a/collectors/ebpf.plugin/ebpf_unittest.h
+++ b/src/collectors/ebpf.plugin/ebpf_unittest.h
diff --git a/collectors/ebpf.plugin/ebpf_vfs.c b/src/collectors/ebpf.plugin/ebpf_vfs.c
index 354901c9c..cb7500aab 100644
--- a/collectors/ebpf.plugin/ebpf_vfs.c
+++ b/src/collectors/ebpf.plugin/ebpf_vfs.c
@@ -1,7 +1,5 @@
// SPDX-License-Identifier: GPL-3.0-or-later
-#include <sys/resource.h>
-
#include "ebpf.h"
#include "ebpf_vfs.h"
@@ -43,6 +41,17 @@ static ebpf_local_maps_t vfs_maps[] = {{.name = "tbl_vfs_pid", .internal_input =
#endif
}};
+struct netdata_static_thread ebpf_read_vfs = {
+ .name = "EBPF_READ_VFS",
+ .config_section = NULL,
+ .config_name = NULL,
+ .env_name = NULL,
+ .enabled = 1,
+ .thread = NULL,
+ .init_routine = NULL,
+ .start_routine = NULL
+};
+
struct config vfs_config = { .first_section = NULL,
.last_section = NULL,
.mutex = NETDATA_MUTEX_INITIALIZER,
@@ -60,10 +69,6 @@ netdata_ebpf_targets_t vfs_targets[] = { {.name = "vfs_write", .mode = EBPF_LOAD
{.name = "release_task", .mode = EBPF_LOAD_TRAMPOLINE},
{.name = NULL, .mode = EBPF_LOAD_TRAMPOLINE}};
-#ifdef NETDATA_DEV_MODE
-int vfs_disable_priority;
-#endif
-
#ifdef LIBBPF_MAJOR_VERSION
/**
* Disable probe
@@ -90,7 +95,6 @@ static void ebpf_vfs_disable_probes(struct vfs_bpf *obj)
bpf_program__set_autoload(obj->progs.netdata_vfs_open_kretprobe, false);
bpf_program__set_autoload(obj->progs.netdata_vfs_create_kprobe, false);
bpf_program__set_autoload(obj->progs.netdata_vfs_create_kretprobe, false);
- bpf_program__set_autoload(obj->progs.netdata_vfs_release_task_kprobe, false);
}
/*
@@ -116,7 +120,6 @@ static void ebpf_vfs_disable_trampoline(struct vfs_bpf *obj)
bpf_program__set_autoload(obj->progs.netdata_vfs_open_fentry, false);
bpf_program__set_autoload(obj->progs.netdata_vfs_open_fexit, false);
bpf_program__set_autoload(obj->progs.netdata_vfs_create_fentry, false);
- bpf_program__set_autoload(obj->progs.netdata_vfs_release_task_fentry, false);
}
/**
@@ -155,8 +158,6 @@ static void ebpf_vfs_set_trampoline_target(struct vfs_bpf *obj)
bpf_program__set_attach_target(obj->progs.netdata_vfs_open_fexit, 0, vfs_targets[NETDATA_EBPF_VFS_OPEN].name);
bpf_program__set_attach_target(obj->progs.netdata_vfs_create_fentry, 0, vfs_targets[NETDATA_EBPF_VFS_CREATE].name);
-
- bpf_program__set_attach_target(obj->progs.netdata_vfs_release_task_fentry, 0, EBPF_COMMON_FNCT_CLEAN_UP);
}
/**
@@ -172,7 +173,7 @@ static int ebpf_vfs_attach_probe(struct vfs_bpf *obj)
{
obj->links.netdata_vfs_write_kprobe = bpf_program__attach_kprobe(obj->progs.netdata_vfs_write_kprobe, false,
vfs_targets[NETDATA_EBPF_VFS_WRITE].name);
- int ret = libbpf_get_error(obj->links.netdata_vfs_write_kprobe);
+ long ret = libbpf_get_error(obj->links.netdata_vfs_write_kprobe);
if (ret)
return -1;
@@ -302,13 +303,6 @@ static int ebpf_vfs_attach_probe(struct vfs_bpf *obj)
if (ret)
return -1;
- obj->links.netdata_vfs_release_task_kprobe = bpf_program__attach_kprobe(obj->progs.netdata_vfs_release_task_fentry,
- true,
- EBPF_COMMON_FNCT_CLEAN_UP);
- ret = libbpf_get_error(obj->links.netdata_vfs_release_task_kprobe);
- if (ret)
- return -1;
-
return 0;
}
@@ -345,19 +339,6 @@ static void ebpf_vfs_set_hash_tables(struct vfs_bpf *obj)
}
/**
- * Disable Release Task
- *
- * Disable release task when apps is not enabled.
- *
- * @param obj is the main structure for bpf objects.
- */
-static void ebpf_vfs_disable_release_task(struct vfs_bpf *obj)
-{
- bpf_program__set_autoload(obj->progs.netdata_vfs_release_task_fentry, false);
- bpf_program__set_autoload(obj->progs.netdata_vfs_release_task_kprobe, false);
-}
-
-/**
* Load and attach
*
* Load and attach the eBPF code in kernel.
@@ -382,9 +363,6 @@ static inline int ebpf_vfs_load_and_attach(struct vfs_bpf *obj, ebpf_module_t *e
ebpf_vfs_adjust_map(obj, em);
- if (!em->apps_charts && !em->cgroup_charts)
- ebpf_vfs_disable_release_task(obj);
-
int ret = vfs_bpf__load(obj);
if (ret) {
return ret;
@@ -416,25 +394,25 @@ static void ebpf_obsolete_specific_vfs_charts(char *type, ebpf_module_t *em);
*
* @param em a pointer to `struct ebpf_module`
*/
-static void ebpf_obsolete_vfs_services(ebpf_module_t *em)
+static void ebpf_obsolete_vfs_services(ebpf_module_t *em, char *id)
{
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_FILE_DELETED,
- "",
"Files deleted",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20065,
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS,
- "",
"Write to disk",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20066,
@@ -442,11 +420,11 @@ static void ebpf_obsolete_vfs_services(ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS_ERROR,
- "",
"Fails to write",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20067,
@@ -454,11 +432,11 @@ static void ebpf_obsolete_vfs_services(ebpf_module_t *em)
}
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_READ_CALLS,
- "",
"Read from disk",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20068,
@@ -466,11 +444,11 @@ static void ebpf_obsolete_vfs_services(ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_READ_CALLS_ERROR,
- "",
"Fails to read",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20069,
@@ -478,33 +456,33 @@ static void ebpf_obsolete_vfs_services(ebpf_module_t *em)
}
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_WRITE_BYTES,
- "",
"Bytes written on disk",
EBPF_COMMON_DIMENSION_BYTES,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20070,
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_READ_BYTES,
- "",
"Bytes read from disk",
EBPF_COMMON_DIMENSION_BYTES,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20071,
em->update_every);
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_FSYNC,
- "",
"Calls to vfs_fsync.",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20072,
@@ -512,22 +490,22 @@ static void ebpf_obsolete_vfs_services(ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_FSYNC_CALLS_ERROR,
- "",
"Sync error",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20073,
em->update_every);
}
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_OPEN,
- "",
"Calls to vfs_open.",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20074,
@@ -535,11 +513,11 @@ static void ebpf_obsolete_vfs_services(ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_OPEN_CALLS_ERROR,
- "",
"Open error",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20075,
@@ -547,11 +525,11 @@ static void ebpf_obsolete_vfs_services(ebpf_module_t *em)
}
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_CREATE,
- "",
"Calls to vfs_create.",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20076,
@@ -559,11 +537,11 @@ static void ebpf_obsolete_vfs_services(ebpf_module_t *em)
if (em->mode < MODE_ENTRY) {
ebpf_write_chart_obsolete(NETDATA_SERVICE_FAMILY,
+ id,
NETDATA_SYSCALL_APPS_VFS_CREATE_CALLS_ERROR,
- "",
"Create error",
EBPF_COMMON_DIMENSION_CALL,
- NETDATA_VFS_CGROUP_GROUP,
+ NETDATA_VFS_GROUP,
NETDATA_EBPF_CHART_TYPE_STACKED,
NULL,
20077,
@@ -581,12 +559,13 @@ static void ebpf_obsolete_vfs_services(ebpf_module_t *em)
static inline void ebpf_obsolete_vfs_cgroup_charts(ebpf_module_t *em) {
pthread_mutex_lock(&mutex_cgroup_shm);
- ebpf_obsolete_vfs_services(em);
-
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (ect->systemd)
+ if (ect->systemd) {
+ ebpf_obsolete_vfs_services(em, ect->name);
+
continue;
+ }
ebpf_obsolete_specific_vfs_charts(ect->name, em);
}
@@ -605,6 +584,7 @@ void ebpf_obsolete_vfs_apps_charts(struct ebpf_module *em)
int order = 20275;
struct ebpf_target *w;
int update_every = em->update_every;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = apps_groups_root_target; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_VFS_IDX))))
continue;
@@ -763,6 +743,7 @@ void ebpf_obsolete_vfs_apps_charts(struct ebpf_module *em)
}
w->charts_created &= ~(1<<EBPF_MODULE_VFS_IDX);
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -904,6 +885,9 @@ static void ebpf_vfs_exit(void *ptr)
{
ebpf_module_t *em = (ebpf_module_t *)ptr;
+ if (ebpf_read_vfs.thread)
+ netdata_thread_cancel(*ebpf_read_vfs.thread);
+
if (em->enabled == NETDATA_THREAD_EBPF_FUNCTION_RUNNING) {
pthread_mutex_lock(&lock);
if (em->cgroup_charts) {
@@ -917,11 +901,6 @@ static void ebpf_vfs_exit(void *ptr)
ebpf_obsolete_vfs_global(em);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_vfs_pid)
- ebpf_statistic_obsolete_aral_chart(em, vfs_disable_priority);
-#endif
-
fflush(stdout);
pthread_mutex_unlock(&lock);
}
@@ -1068,8 +1047,9 @@ static void ebpf_vfs_sum_pids(netdata_publish_vfs_t *vfs, struct ebpf_pid_on_tar
while (root) {
int32_t pid = root->pid;
- netdata_publish_vfs_t *w = vfs_pid[pid];
- if (w) {
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ netdata_publish_vfs_t *w = &local_pid->vfs;
accumulator.write_call += w->write_call;
accumulator.writev_call += w->writev_call;
accumulator.read_call += w->read_call;
@@ -1130,12 +1110,11 @@ static void ebpf_vfs_sum_pids(netdata_publish_vfs_t *vfs, struct ebpf_pid_on_tar
void ebpf_vfs_send_apps_data(ebpf_module_t *em, struct ebpf_target *root)
{
struct ebpf_target *w;
+ pthread_mutex_lock(&collect_data_mutex);
for (w = root; w; w = w->next) {
if (unlikely(!(w->charts_created & (1<<EBPF_MODULE_VFS_IDX))))
continue;
- ebpf_vfs_sum_pids(&w->vfs, w->root_pid);
-
ebpf_write_begin_chart(NETDATA_APP_FAMILY, w->clean_name, "_ebpf_call_vfs_unlink");
write_chart_dimension("calls", w->vfs.unlink_call);
ebpf_write_end_chart();
@@ -1198,6 +1177,7 @@ void ebpf_vfs_send_apps_data(ebpf_module_t *em, struct ebpf_target *root)
ebpf_write_end_chart();
}
}
+ pthread_mutex_unlock(&collect_data_mutex);
}
/**
@@ -1234,52 +1214,41 @@ static void vfs_apps_accumulator(netdata_publish_vfs_t *out, int maps_per_core)
}
/**
- * Fill PID
- *
- * Fill PID structures
- *
- * @param current_pid pid that we are collecting data
- * @param out values read from hash tables;
- */
-static void vfs_fill_pid(uint32_t current_pid, netdata_publish_vfs_t *publish)
-{
- netdata_publish_vfs_t *curr = vfs_pid[current_pid];
- if (!curr) {
- curr = ebpf_vfs_get();
- vfs_pid[current_pid] = curr;
- }
-
- memcpy(curr, &publish[0], sizeof(netdata_publish_vfs_t));
-}
-
-/**
* Read the hash table and store data to allocated vectors.
*/
-static void ebpf_vfs_read_apps(int maps_per_core)
+static void ebpf_vfs_read_apps(int maps_per_core, int max_period)
{
- struct ebpf_pid_stat *pids = ebpf_root_of_pids;
netdata_publish_vfs_t *vv = vfs_vector;
int fd = vfs_maps[NETDATA_VFS_PID].map_fd;
size_t length = sizeof(netdata_publish_vfs_t);
if (maps_per_core)
length *= ebpf_nprocs;
- while (pids) {
- uint32_t key = pids->pid;
-
+ uint32_t key = 0, next_key = 0;
+ while (bpf_map_get_next_key(fd, &key, &next_key) == 0) {
if (bpf_map_lookup_elem(fd, &key, vv)) {
- pids = pids->next;
- continue;
+ goto end_vfs_loop;
}
vfs_apps_accumulator(vv, maps_per_core);
- vfs_fill_pid(key, vv);
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(key, vv->tgid);
+ if (!local_pid)
+ goto end_vfs_loop;
+
+ netdata_publish_vfs_t *publish = &local_pid->vfs;
+ if (!publish->ct || publish->ct != vv->ct) {
+ memcpy(publish, vv, sizeof(netdata_publish_vfs_t));
+ local_pid->not_updated = 0;
+ } else if (++local_pid->not_updated >= max_period){
+ bpf_map_delete_elem(fd, &key);
+ local_pid->not_updated = 0;
+ }
+end_vfs_loop:
// We are cleaning to avoid passing data read from one process to other.
memset(vv, 0, length);
-
- pids = pids->next;
+ key = next_key;
}
}
@@ -1290,32 +1259,20 @@ static void ebpf_vfs_read_apps(int maps_per_core)
*
* @param maps_per_core do I need to read all cores?
*/
-static void read_update_vfs_cgroup(int maps_per_core)
+static void read_update_vfs_cgroup()
{
ebpf_cgroup_target_t *ect ;
- netdata_publish_vfs_t *vv = vfs_vector;
- int fd = vfs_maps[NETDATA_VFS_PID].map_fd;
- size_t length = sizeof(netdata_publish_vfs_t);
- if (maps_per_core)
- length *= ebpf_nprocs;
-
pthread_mutex_lock(&mutex_cgroup_shm);
for (ect = ebpf_cgroup_pids; ect; ect = ect->next) {
struct pid_on_target2 *pids;
for (pids = ect->pids; pids; pids = pids->next) {
int pid = pids->pid;
netdata_publish_vfs_t *out = &pids->vfs;
- if (likely(vfs_pid) && vfs_pid[pid]) {
- netdata_publish_vfs_t *in = vfs_pid[pid];
+ ebpf_pid_stat_t *local_pid = ebpf_get_pid_entry(pid, 0);
+ if (local_pid) {
+ netdata_publish_vfs_t *in = &local_pid->vfs;
memcpy(out, in, sizeof(netdata_publish_vfs_t));
- } else {
- memset(vv, 0, length);
- if (!bpf_map_lookup_elem(fd, &pid, vv)) {
- vfs_apps_accumulator(vv, maps_per_core);
-
- memcpy(out, vv, sizeof(netdata_publish_vfs_t));
- }
}
}
}
@@ -1400,88 +1357,88 @@ static void ebpf_vfs_sum_cgroup_pids(netdata_publish_vfs_t *vfs, struct pid_on_t
static void ebpf_create_specific_vfs_charts(char *type, ebpf_module_t *em)
{
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_FILE_DELETED,"Files deleted",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_UNLINK_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_UNLINK_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5500,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_UNLINK],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS, "Write to disk",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_WRITE_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_WRITE_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5501,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_WRITE],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
if (em->mode < MODE_ENTRY) {
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS_ERROR, "Fails to write",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_WRITE_ERROR_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_WRITE_ERROR_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5502,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_WRITE],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
}
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_READ_CALLS, "Read from disk",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_READ_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_READ_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5503,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_READ],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
if (em->mode < MODE_ENTRY) {
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_READ_CALLS_ERROR, "Fails to read",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_READ_ERROR_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_READ_ERROR_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5504,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_READ],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
}
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_WRITE_BYTES, "Bytes written on disk",
- EBPF_COMMON_DIMENSION_BYTES, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_WRITE_BYTES_CONTEXT,
+ EBPF_COMMON_DIMENSION_BYTES, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_WRITE_BYTES_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5505,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_WRITE],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_READ_BYTES, "Bytes read from disk",
- EBPF_COMMON_DIMENSION_BYTES, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_READ_BYTES_CONTEXT,
+ EBPF_COMMON_DIMENSION_BYTES, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_READ_BYTES_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5506,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_READ],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_FSYNC, "Calls to vfs_fsync.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_FSYNC_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_FSYNC_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5507,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_FSYNC],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
if (em->mode < MODE_ENTRY) {
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_FSYNC_CALLS_ERROR, "Sync error",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_FSYNC_ERROR_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_FSYNC_ERROR_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5508,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_FSYNC],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
}
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_OPEN, "Calls to vfs_open.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_OPEN_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_OPEN_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5509,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_OPEN],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
if (em->mode < MODE_ENTRY) {
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_OPEN_CALLS_ERROR, "Open error",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_OPEN_ERROR_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_OPEN_ERROR_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5510,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_OPEN],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
}
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_CREATE, "Calls to vfs_create.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_CREATE_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_CREATE_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5511,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_CREATE],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
if (em->mode < MODE_ENTRY) {
ebpf_create_chart(type, NETDATA_SYSCALL_APPS_VFS_CREATE_CALLS_ERROR, "Create error",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP, NETDATA_CGROUP_VFS_CREATE_ERROR_CONTEXT,
+ EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_GROUP, NETDATA_CGROUP_VFS_CREATE_ERROR_CONTEXT,
NETDATA_EBPF_CHART_TYPE_LINE, NETDATA_CHART_PRIO_CGROUPS_CONTAINERS + 5512,
ebpf_create_global_dimension, &vfs_publish_aggregated[NETDATA_KEY_PUBLISH_VFS_CREATE],
1, em->update_every, NETDATA_EBPF_MODULE_NAME_VFS);
@@ -1668,93 +1625,235 @@ static void ebpf_send_specific_vfs_data(char *type, netdata_publish_vfs_t *value
**/
static void ebpf_create_systemd_vfs_charts(ebpf_module_t *em)
{
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_FILE_DELETED, "Files deleted",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20065,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_UNLINK_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
-
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS, "Write to disk",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20066,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_WRITE_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
+ static ebpf_systemd_args_t data_vfs_unlink = {
+ .title = "Files deleted",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20065,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_UNLINK_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_FILE_DELETED,
+ .dimension = "calls"
+ };
- if (em->mode < MODE_ENTRY) {
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS_ERROR, "Fails to write",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20067,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SYSTEMD_VFS_WRITE_ERROR_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
- }
+ static ebpf_systemd_args_t data_vfs_write = {
+ .title = "Write to disk",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20066,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_WRITE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_READ_CALLS, "Read from disk",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20068,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_READ_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
+ static ebpf_systemd_args_t data_vfs_write_err = {
+ .title = "Fails to write",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20067,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_WRITE_ERROR_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS_ERROR,
+ .dimension = "calls"
+ };
- if (em->mode < MODE_ENTRY) {
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_READ_CALLS_ERROR, "Fails to read",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20069,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX],
- NETDATA_SYSTEMD_VFS_READ_ERROR_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
- }
+ static ebpf_systemd_args_t data_vfs_read = {
+ .title = "Read from disk",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20068,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_READ_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_READ_CALLS,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_WRITE_BYTES, "Bytes written on disk",
- EBPF_COMMON_DIMENSION_BYTES, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20070,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_WRITE_BYTES_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
+ static ebpf_systemd_args_t data_vfs_read_err = {
+ .title = "Fails to read",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20069,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_READ_ERROR_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_READ_CALLS_ERROR,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_READ_BYTES, "Bytes read from disk",
- EBPF_COMMON_DIMENSION_BYTES, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20071,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_READ_BYTES_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
+ static ebpf_systemd_args_t data_vfs_write_bytes = {
+ .title = "Bytes written on disk",
+ .units = EBPF_COMMON_DIMENSION_BYTES,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20070,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_WRITE_BYTES_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_WRITE_BYTES,
+ .dimension = "bytes"
+ };
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_FSYNC, "Calls to vfs_fsync.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20072,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_FSYNC_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
+ static ebpf_systemd_args_t data_vfs_read_bytes = {
+ .title = "Bytes read from disk",
+ .units = EBPF_COMMON_DIMENSION_BYTES,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20071,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_READ_BYTES_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_READ_BYTES,
+ .dimension = "bytes"
+ };
- if (em->mode < MODE_ENTRY) {
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_FSYNC_CALLS_ERROR, "Sync error",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20073,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_FSYNC_ERROR_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
- }
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_OPEN, "Calls to vfs_open.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20074,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_OPEN_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
+ static ebpf_systemd_args_t data_vfs_fsync = {
+ .title = "Calls to vfs_fsync.",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20072,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_FSYNC_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_FSYNC,
+ .dimension = "calls"
+ };
- if (em->mode < MODE_ENTRY) {
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_OPEN_CALLS_ERROR, "Open error",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20075,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_OPEN_ERROR_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
- }
+ static ebpf_systemd_args_t data_vfs_fsync_err = {
+ .title = "Sync error",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20073,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_FSYNC_ERROR_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_FSYNC_CALLS_ERROR,
+ .dimension = "calls"
+ };
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_CREATE, "Calls to vfs_create.",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20076,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_CREATE_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
+ static ebpf_systemd_args_t data_vfs_open = {
+ .title = "Calls to vfs_open.",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20074,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_OPEN_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_OPEN,
+ .dimension = "calls"
+ };
- if (em->mode < MODE_ENTRY) {
- ebpf_create_charts_on_systemd(NETDATA_SYSCALL_APPS_VFS_CREATE_CALLS_ERROR, "Create error",
- EBPF_COMMON_DIMENSION_CALL, NETDATA_VFS_CGROUP_GROUP,
- NETDATA_EBPF_CHART_TYPE_STACKED, 20077,
- ebpf_algorithms[NETDATA_EBPF_INCREMENTAL_IDX], NETDATA_SYSTEMD_VFS_CREATE_ERROR_CONTEXT,
- NETDATA_EBPF_MODULE_NAME_VFS, em->update_every);
+ static ebpf_systemd_args_t data_vfs_open_err = {
+ .title = "Open error",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20075,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_OPEN_ERROR_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_OPEN_CALLS_ERROR,
+ .dimension = "calls"
+ };
+
+ static ebpf_systemd_args_t data_vfs_create = {
+ .title = "Calls to vfs_create.",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20076,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_CREATE_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_CREATE,
+ .dimension = "calls"
+ };
+
+ static ebpf_systemd_args_t data_vfs_create_err = {
+ .title = "Create error",
+ .units = EBPF_COMMON_DIMENSION_CALL,
+ .family = NETDATA_VFS_GROUP,
+ .charttype = NETDATA_EBPF_CHART_TYPE_STACKED,
+ .order = 20077,
+ .algorithm = EBPF_CHART_ALGORITHM_INCREMENTAL,
+ .context = NETDATA_SYSTEMD_VFS_CREATE_ERROR_CONTEXT,
+ .module = NETDATA_EBPF_MODULE_NAME_VFS,
+ .update_every = 0,
+ .suffix = NETDATA_SYSCALL_APPS_VFS_CREATE_CALLS_ERROR,
+ .dimension = "calls"
+ };
+
+ if (!data_vfs_create.update_every)
+ data_vfs_unlink.update_every = data_vfs_write.update_every = data_vfs_write_err.update_every =
+ data_vfs_read.update_every = data_vfs_read_err.update_every = data_vfs_write_bytes.update_every =
+ data_vfs_read_bytes.update_every = data_vfs_fsync.update_every = data_vfs_fsync_err.update_every =
+ data_vfs_open.update_every = data_vfs_open_err.update_every = data_vfs_create.update_every =
+ data_vfs_create_err.update_every = em->update_every;
+
+ ebpf_cgroup_target_t *w;
+ for (w = ebpf_cgroup_pids; w ; w = w->next) {
+ if (unlikely(!w->systemd || w->flags & NETDATA_EBPF_SERVICES_HAS_VFS_CHART))
+ continue;
+
+ data_vfs_unlink.id = data_vfs_write.id = data_vfs_write_err.id =
+ data_vfs_read.id = data_vfs_read_err.id = data_vfs_write_bytes.id = data_vfs_read_bytes.id =
+ data_vfs_fsync.id = data_vfs_fsync_err.id = data_vfs_open.id =
+ data_vfs_open_err.id = data_vfs_create.id = data_vfs_create_err.id = w->name;
+ ebpf_create_charts_on_systemd(&data_vfs_unlink);
+
+ ebpf_create_charts_on_systemd(&data_vfs_write);
+ if (em->mode < MODE_ENTRY) {
+ ebpf_create_charts_on_systemd(&data_vfs_write_err);
+ }
+
+ ebpf_create_charts_on_systemd(&data_vfs_read);
+ if (em->mode < MODE_ENTRY) {
+ ebpf_create_charts_on_systemd(&data_vfs_read_err);
+ }
+
+ ebpf_create_charts_on_systemd(&data_vfs_write_bytes);
+ ebpf_create_charts_on_systemd(&data_vfs_read_bytes);
+
+ ebpf_create_charts_on_systemd(&data_vfs_fsync);
+ if (em->mode < MODE_ENTRY) {
+ ebpf_create_charts_on_systemd(&data_vfs_fsync_err);
+ }
+
+ ebpf_create_charts_on_systemd(&data_vfs_open);
+ if (em->mode < MODE_ENTRY) {
+ ebpf_create_charts_on_systemd(&data_vfs_open_err);
+ }
+
+ ebpf_create_charts_on_systemd(&data_vfs_create);
+ if (em->mode < MODE_ENTRY) {
+ ebpf_create_charts_on_systemd(&data_vfs_create_err);
+ }
+
+ w->flags |= NETDATA_EBPF_SERVICES_HAS_VFS_CHART;
}
}
@@ -1768,124 +1867,78 @@ static void ebpf_create_systemd_vfs_charts(ebpf_module_t *em)
static void ebpf_send_systemd_vfs_charts(ebpf_module_t *em)
{
ebpf_cgroup_target_t *ect;
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_FILE_DELETED, "");
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.unlink_call);
+ if (unlikely(!(ect->flags & NETDATA_EBPF_SERVICES_HAS_VFS_CHART)) ) {
+ continue;
}
- }
- ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.write_call +
- ect->publish_systemd_vfs.writev_call);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_FILE_DELETED);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.unlink_call);
+ ebpf_write_end_chart();
- if (em->mode < MODE_ENTRY) {
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS_ERROR, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.write_err +
- ect->publish_systemd_vfs.writev_err);
- }
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.write_call +
+ ect->publish_systemd_vfs.writev_call);
ebpf_write_end_chart();
- }
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_READ_CALLS, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.read_call +
- ect->publish_systemd_vfs.readv_call);
+ if (em->mode < MODE_ENTRY) {
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_WRITE_CALLS_ERROR);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.write_err +
+ ect->publish_systemd_vfs.writev_err);
+ ebpf_write_end_chart();
}
- }
- ebpf_write_end_chart();
- if (em->mode < MODE_ENTRY) {
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_READ_CALLS_ERROR, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.read_err +
- ect->publish_systemd_vfs.readv_err);
- }
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_READ_CALLS);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.read_call +
+ ect->publish_systemd_vfs.readv_call);
ebpf_write_end_chart();
- }
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_WRITE_BYTES, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.write_bytes +
- ect->publish_systemd_vfs.writev_bytes);
+ if (em->mode < MODE_ENTRY) {
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_READ_CALLS_ERROR);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.read_err +
+ ect->publish_systemd_vfs.readv_err);
+ ebpf_write_end_chart();
}
- }
- ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_READ_BYTES, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.read_bytes +
- ect->publish_systemd_vfs.readv_bytes);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_WRITE_BYTES);
+ write_chart_dimension("bytes", ect->publish_systemd_vfs.write_bytes +
+ ect->publish_systemd_vfs.writev_bytes);
+ ebpf_write_end_chart();
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_FSYNC, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.fsync_call);
- }
- }
- ebpf_write_end_chart();
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_READ_BYTES);
+ write_chart_dimension("bytes", ect->publish_systemd_vfs.read_bytes +
+ ect->publish_systemd_vfs.readv_bytes);
+ ebpf_write_end_chart();
- if (em->mode < MODE_ENTRY) {
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_FSYNC_CALLS_ERROR, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.fsync_err);
- }
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_FSYNC);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.fsync_call);
ebpf_write_end_chart();
- }
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_OPEN, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.open_call);
+ if (em->mode < MODE_ENTRY) {
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_FSYNC_CALLS_ERROR);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.fsync_err);
+ ebpf_write_end_chart();
}
- }
- ebpf_write_end_chart();
- if (em->mode < MODE_ENTRY) {
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_OPEN_CALLS_ERROR, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.open_err);
- }
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_OPEN);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.open_call);
ebpf_write_end_chart();
- }
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_CREATE, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.create_call);
+ if (em->mode < MODE_ENTRY) {
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_OPEN_CALLS_ERROR);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.open_err);
+ ebpf_write_end_chart();
}
- }
- ebpf_write_end_chart();
- if (em->mode < MODE_ENTRY) {
- ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, NETDATA_SYSCALL_APPS_VFS_CREATE_CALLS_ERROR, "");
- for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
- if (unlikely(ect->systemd) && unlikely(ect->updated)) {
- write_chart_dimension(ect->name, ect->publish_systemd_vfs.create_err);
- }
- }
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_CREATE);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.create_call);
ebpf_write_end_chart();
+
+ if (em->mode < MODE_ENTRY) {
+ ebpf_write_begin_chart(NETDATA_SERVICE_FAMILY, ect->name, NETDATA_SYSCALL_APPS_VFS_CREATE_CALLS_ERROR);
+ write_chart_dimension("calls", ect->publish_systemd_vfs.create_err);
+ ebpf_write_end_chart();
+ }
}
}
@@ -1896,17 +1949,13 @@ static void ebpf_send_systemd_vfs_charts(ebpf_module_t *em)
*/
static void ebpf_vfs_send_cgroup_data(ebpf_module_t *em)
{
- if (!ebpf_cgroup_pids)
- return;
-
pthread_mutex_lock(&mutex_cgroup_shm);
ebpf_cgroup_target_t *ect;
for (ect = ebpf_cgroup_pids; ect ; ect = ect->next) {
ebpf_vfs_sum_cgroup_pids(&ect->publish_systemd_vfs, ect->pids);
}
- int has_systemd = shm_ebpf_cgroup.header->systemd_enabled;
- if (has_systemd) {
+ if (shm_ebpf_cgroup.header->systemd_enabled) {
if (send_cgroup_chart) {
ebpf_create_systemd_vfs_charts(em);
}
@@ -1936,6 +1985,72 @@ static void ebpf_vfs_send_cgroup_data(ebpf_module_t *em)
}
/**
+ * Resume apps data
+ */
+void ebpf_vfs_resume_apps_data() {
+ struct ebpf_target *w;
+ for (w = apps_groups_root_target; w; w = w->next) {
+ if (unlikely(!(w->charts_created & (1 << EBPF_MODULE_VFS_IDX))))
+ continue;
+
+ ebpf_vfs_sum_pids(&w->vfs, w->root_pid);
+ }
+}
+
+/**
+ * VFS thread
+ *
+ * Thread used to generate charts.
+ *
+ * @param ptr a pointer to `struct ebpf_module`
+ *
+ * @return It always return NULL
+ */
+void *ebpf_read_vfs_thread(void *ptr)
+{
+ heartbeat_t hb;
+ heartbeat_init(&hb);
+
+ ebpf_module_t *em = (ebpf_module_t *)ptr;
+
+ int maps_per_core = em->maps_per_core;
+ int update_every = em->update_every;
+
+ int counter = update_every - 1;
+
+ uint32_t lifetime = em->lifetime;
+ uint32_t running_time = 0;
+ usec_t period = update_every * USEC_PER_SEC;
+ int max_period = update_every * EBPF_CLEANUP_FACTOR;
+ while (!ebpf_plugin_exit && running_time < lifetime) {
+ (void)heartbeat_next(&hb, period);
+ if (ebpf_plugin_exit || ++counter != update_every)
+ continue;
+
+ netdata_thread_disable_cancelability();
+
+ pthread_mutex_lock(&collect_data_mutex);
+ ebpf_vfs_read_apps(maps_per_core, max_period);
+ ebpf_vfs_resume_apps_data();
+ pthread_mutex_unlock(&collect_data_mutex);
+
+ counter = 0;
+
+ pthread_mutex_lock(&ebpf_exit_cleanup);
+ if (running_time && !em->running_time)
+ running_time = update_every;
+ else
+ running_time += update_every;
+
+ em->running_time = running_time;
+ pthread_mutex_unlock(&ebpf_exit_cleanup);
+ netdata_thread_enable_cancelability();
+ }
+
+ return NULL;
+}
+
+/**
* Main loop for this collector.
*
* @param step the number of microseconds used with heart beat
@@ -1961,31 +2076,22 @@ static void vfs_collector(ebpf_module_t *em)
counter = 0;
netdata_apps_integration_flags_t apps = em->apps_charts;
ebpf_vfs_read_global_table(stats, maps_per_core);
- pthread_mutex_lock(&collect_data_mutex);
- if (apps)
- ebpf_vfs_read_apps(maps_per_core);
if (cgroups)
- read_update_vfs_cgroup(maps_per_core);
+ read_update_vfs_cgroup();
pthread_mutex_lock(&lock);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_vfs_pid)
- ebpf_send_data_aral_chart(ebpf_aral_vfs_pid, em);
-#endif
-
ebpf_vfs_send_data(em);
fflush(stdout);
if (apps & NETDATA_EBPF_APPS_FLAG_CHART_CREATED)
ebpf_vfs_send_apps_data(em, apps_groups_root_target);
- if (cgroups)
+ if (cgroups && shm_ebpf_cgroup.header && ebpf_cgroup_pids)
ebpf_vfs_send_cgroup_data(em);
pthread_mutex_unlock(&lock);
- pthread_mutex_unlock(&collect_data_mutex);
pthread_mutex_lock(&ebpf_exit_cleanup);
if (running_time && !em->running_time)
@@ -2412,13 +2518,9 @@ void ebpf_vfs_create_apps_charts(struct ebpf_module *em, void *ptr)
*
* @param apps is apps enabled?
*/
-static void ebpf_vfs_allocate_global_vectors(int apps)
+static void ebpf_vfs_allocate_global_vectors()
{
- if (apps) {
- ebpf_vfs_aral_init();
- vfs_pid = callocz((size_t)pid_max, sizeof(netdata_publish_vfs_t *));
- vfs_vector = callocz(ebpf_nprocs, sizeof(netdata_publish_vfs_t));
- }
+ vfs_vector = callocz(ebpf_nprocs, sizeof(netdata_publish_vfs_t));
memset(vfs_aggregated_data, 0, sizeof(vfs_aggregated_data));
memset(vfs_publish_aggregated, 0, sizeof(vfs_publish_aggregated));
@@ -2484,7 +2586,7 @@ void *ebpf_vfs_thread(void *ptr)
ebpf_update_pid_table(&vfs_maps[NETDATA_VFS_PID], em);
- ebpf_vfs_allocate_global_vectors(em->apps_charts);
+ ebpf_vfs_allocate_global_vectors();
#ifdef LIBBPF_MAJOR_VERSION
ebpf_adjust_thread_load(em, default_btf);
@@ -2505,13 +2607,16 @@ void *ebpf_vfs_thread(void *ptr)
ebpf_create_global_charts(em);
ebpf_update_stats(&plugin_statistics, em);
ebpf_update_kernel_memory_with_vector(&plugin_statistics, em->maps, EBPF_ACTION_STAT_ADD);
-#ifdef NETDATA_DEV_MODE
- if (ebpf_aral_vfs_pid)
- vfs_disable_priority = ebpf_statistic_create_aral_chart(NETDATA_EBPF_VFS_ARAL_NAME, em);
-#endif
pthread_mutex_unlock(&lock);
+ ebpf_read_vfs.thread = mallocz(sizeof(netdata_thread_t));
+ netdata_thread_create(ebpf_read_vfs.thread,
+ ebpf_read_vfs.name,
+ NETDATA_THREAD_OPTION_DEFAULT,
+ ebpf_read_vfs_thread,
+ em);
+
vfs_collector(em);
endvfs:
diff --git a/collectors/ebpf.plugin/ebpf_vfs.h b/src/collectors/ebpf.plugin/ebpf_vfs.h
index 8fe12a7eb..89364a2b5 100644
--- a/collectors/ebpf.plugin/ebpf_vfs.h
+++ b/src/collectors/ebpf.plugin/ebpf_vfs.h
@@ -39,7 +39,6 @@
// Group used on Dashboard
#define NETDATA_VFS_GROUP "vfs"
-#define NETDATA_VFS_CGROUP_GROUP "vfs (eBPF)"
// Contexts
#define NETDATA_CGROUP_VFS_UNLINK_CONTEXT "cgroup.vfs_unlink"
@@ -56,27 +55,29 @@
#define NETDATA_CGROUP_VFS_FSYNC_CONTEXT "cgroup.vfs_fsync"
#define NETDATA_CGROUP_VFS_FSYNC_ERROR_CONTEXT "cgroup.vfs_fsync_error"
-#define NETDATA_SYSTEMD_VFS_UNLINK_CONTEXT "services.vfs_unlink"
-#define NETDATA_SYSTEMD_VFS_WRITE_CONTEXT "services.vfs_write"
-#define NETDATA_SYSTEMD_VFS_WRITE_ERROR_CONTEXT "services.vfs_write_error"
-#define NETDATA_SYSTEMD_VFS_READ_CONTEXT "services.vfs_read"
-#define NETDATA_SYSTEMD_VFS_READ_ERROR_CONTEXT "services.vfs_read_error"
-#define NETDATA_SYSTEMD_VFS_WRITE_BYTES_CONTEXT "services.vfs_write_bytes"
-#define NETDATA_SYSTEMD_VFS_READ_BYTES_CONTEXT "services.vfs_read_bytes"
-#define NETDATA_SYSTEMD_VFS_CREATE_CONTEXT "services.vfs_create"
-#define NETDATA_SYSTEMD_VFS_CREATE_ERROR_CONTEXT "services.vfs_create_error"
-#define NETDATA_SYSTEMD_VFS_OPEN_CONTEXT "services.vfs_open"
-#define NETDATA_SYSTEMD_VFS_OPEN_ERROR_CONTEXT "services.vfs_open_error"
-#define NETDATA_SYSTEMD_VFS_FSYNC_CONTEXT "services.vfs_fsync"
-#define NETDATA_SYSTEMD_VFS_FSYNC_ERROR_CONTEXT "services.vfs_fsync_error"
+#define NETDATA_SYSTEMD_VFS_UNLINK_CONTEXT "systemd.services.vfs_unlink"
+#define NETDATA_SYSTEMD_VFS_WRITE_CONTEXT "systemd.services.vfs_write"
+#define NETDATA_SYSTEMD_VFS_WRITE_ERROR_CONTEXT "systemd.services.vfs_write_error"
+#define NETDATA_SYSTEMD_VFS_READ_CONTEXT "systemd.services.vfs_read"
+#define NETDATA_SYSTEMD_VFS_READ_ERROR_CONTEXT "systemd.services.vfs_read_error"
+#define NETDATA_SYSTEMD_VFS_WRITE_BYTES_CONTEXT "systemd.services.vfs_write_bytes"
+#define NETDATA_SYSTEMD_VFS_READ_BYTES_CONTEXT "systemd.services.vfs_read_bytes"
+#define NETDATA_SYSTEMD_VFS_CREATE_CONTEXT "systemd.services.vfs_create"
+#define NETDATA_SYSTEMD_VFS_CREATE_ERROR_CONTEXT "systemd.services.vfs_create_error"
+#define NETDATA_SYSTEMD_VFS_OPEN_CONTEXT "systemd.services.vfs_open"
+#define NETDATA_SYSTEMD_VFS_OPEN_ERROR_CONTEXT "systemd.services.vfs_open_error"
+#define NETDATA_SYSTEMD_VFS_FSYNC_CONTEXT "systemd.services.vfs_fsync"
+#define NETDATA_SYSTEMD_VFS_FSYNC_ERROR_CONTEXT "systemd.services.vfs_fsync_error"
// ARAL name
#define NETDATA_EBPF_VFS_ARAL_NAME "ebpf_vfs"
typedef struct netdata_publish_vfs {
- uint64_t pid_tgid;
- uint32_t pid;
- uint32_t pad;
+ uint64_t ct;
+ uint32_t tgid;
+ uint32_t uid;
+ uint32_t gid;
+ char name[TASK_COMM_LEN];
//Counter
uint32_t write_call;
diff --git a/collectors/ebpf.plugin/integrations/ebpf_cachestat.md b/src/collectors/ebpf.plugin/integrations/ebpf_cachestat.md
index 5bf0a3774..f56cd9533 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_cachestat.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_cachestat.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_cachestat.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_cachestat.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF Cachestat"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -146,7 +146,7 @@ The configuration file name for this integration is `ebpf.d/cachestat.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_dcstat.md b/src/collectors/ebpf.plugin/integrations/ebpf_dcstat.md
index 4c5719026..97f562338 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_dcstat.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_dcstat.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_dcstat.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_dcstat.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF DCstat"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -144,7 +144,7 @@ The configuration file name for this integration is `ebpf.d/dcstat.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_disk.md b/src/collectors/ebpf.plugin/integrations/ebpf_disk.md
index 557da125d..5ea848d5c 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_disk.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_disk.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_disk.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_disk.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF Disk"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -110,7 +110,7 @@ The configuration file name for this integration is `ebpf.d/disk.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_filedescriptor.md b/src/collectors/ebpf.plugin/integrations/ebpf_filedescriptor.md
index 23f5bd26e..684c84efe 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_filedescriptor.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_filedescriptor.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_filedescriptor.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_filedescriptor.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF Filedescriptor"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -144,7 +144,7 @@ The configuration file name for this integration is `ebpf.d/fd.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_filesystem.md b/src/collectors/ebpf.plugin/integrations/ebpf_filesystem.md
index 7a1bb832b..398e247c4 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_filesystem.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_filesystem.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_filesystem.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_filesystem.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF Filesystem"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -131,7 +131,7 @@ The configuration file name for this integration is `ebpf.d/filesystem.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_hardirq.md b/src/collectors/ebpf.plugin/integrations/ebpf_hardirq.md
index f9b529624..c59240bf3 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_hardirq.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_hardirq.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_hardirq.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_hardirq.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF Hardirq"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -110,7 +110,7 @@ The configuration file name for this integration is `ebpf.d/hardirq.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_mdflush.md b/src/collectors/ebpf.plugin/integrations/ebpf_mdflush.md
index 0081b7d83..ce33f9264 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_mdflush.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_mdflush.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_mdflush.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_mdflush.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF MDflush"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -105,7 +105,7 @@ The configuration file name for this integration is `ebpf.d/mdflush.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_mount.md b/src/collectors/ebpf.plugin/integrations/ebpf_mount.md
index d19e57809..f3bd8dff8 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_mount.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_mount.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_mount.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_mount.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF Mount"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -111,7 +111,7 @@ The configuration file name for this integration is `ebpf.d/mount.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_oomkill.md b/src/collectors/ebpf.plugin/integrations/ebpf_oomkill.md
index 897cddfac..ff912ae7b 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_oomkill.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_oomkill.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_oomkill.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_oomkill.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF OOMkill"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -127,7 +127,7 @@ The configuration file name for this integration is `ebpf.d/oomkill.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_process.md b/src/collectors/ebpf.plugin/integrations/ebpf_process.md
index 109890139..df9b5c3d2 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_process.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_process.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_process.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_process.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF Process"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/ebpf.plugin/integrations/ebpf_processes.md b/src/collectors/ebpf.plugin/integrations/ebpf_processes.md
index 62542359a..032e1f648 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_processes.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_processes.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_processes.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_processes.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF Processes"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -154,7 +154,7 @@ The configuration file name for this integration is `ebpf.d/process.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_shm.md b/src/collectors/ebpf.plugin/integrations/ebpf_shm.md
index ffa05c770..5ac0c501f 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_shm.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_shm.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_shm.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_shm.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF SHM"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -148,7 +148,7 @@ The configuration file name for this integration is `ebpf.d/shm.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_socket.md b/src/collectors/ebpf.plugin/integrations/ebpf_socket.md
index dc7a7d07b..8a0286b47 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_socket.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_socket.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_socket.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_socket.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF Socket"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -165,7 +165,7 @@ The configuration file name for this integration is `ebpf.d/network.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_softirq.md b/src/collectors/ebpf.plugin/integrations/ebpf_softirq.md
index 6a4312c6e..8429aac29 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_softirq.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_softirq.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_softirq.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_softirq.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF SoftIRQ"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -110,7 +110,7 @@ The configuration file name for this integration is `ebpf.d/softirq.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_swap.md b/src/collectors/ebpf.plugin/integrations/ebpf_swap.md
index ce2423f8d..57f435170 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_swap.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_swap.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_swap.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_swap.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF SWAP"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -137,7 +137,7 @@ The configuration file name for this integration is `ebpf.d/swap.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_sync.md b/src/collectors/ebpf.plugin/integrations/ebpf_sync.md
index 6f6c246a7..dd8a399a2 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_sync.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_sync.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_sync.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_sync.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF Sync"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -80,7 +80,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ sync_freq ](https://github.com/netdata/netdata/blob/master/health/health.d/synchronization.conf) | mem.sync | number of sync() system calls. Every call causes all pending modifications to filesystem metadata and cached file data to be written to the underlying filesystems. |
+| [ sync_freq ](https://github.com/netdata/netdata/blob/master/src/health/health.d/synchronization.conf) | mem.sync | number of sync() system calls. Every call causes all pending modifications to filesystem metadata and cached file data to be written to the underlying filesystems. |
## Setup
@@ -118,7 +118,7 @@ The configuration file name for this integration is `ebpf.d/sync.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ebpf.plugin/integrations/ebpf_vfs.md b/src/collectors/ebpf.plugin/integrations/ebpf_vfs.md
index 4b824e975..dd182115d 100644
--- a/collectors/ebpf.plugin/integrations/ebpf_vfs.md
+++ b/src/collectors/ebpf.plugin/integrations/ebpf_vfs.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/integrations/ebpf_vfs.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ebpf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/integrations/ebpf_vfs.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ebpf.plugin/metadata.yaml"
sidebar_label: "eBPF VFS"
learn_status: "Published"
-learn_rel_path: "Data Collection/eBPF"
+learn_rel_path: "Collecting Metrics/eBPF"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -179,7 +179,7 @@ The configuration file name for this integration is `ebpf.d/vfs.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/freebsd_devstat.c b/src/collectors/freebsd.plugin/freebsd_devstat.c
index ca6048a16..ca6048a16 100644
--- a/collectors/freebsd.plugin/freebsd_devstat.c
+++ b/src/collectors/freebsd.plugin/freebsd_devstat.c
diff --git a/collectors/freebsd.plugin/freebsd_getifaddrs.c b/src/collectors/freebsd.plugin/freebsd_getifaddrs.c
index 36be68422..36be68422 100644
--- a/collectors/freebsd.plugin/freebsd_getifaddrs.c
+++ b/src/collectors/freebsd.plugin/freebsd_getifaddrs.c
diff --git a/collectors/freebsd.plugin/freebsd_getmntinfo.c b/src/collectors/freebsd.plugin/freebsd_getmntinfo.c
index d55eb3d4a..d55eb3d4a 100644
--- a/collectors/freebsd.plugin/freebsd_getmntinfo.c
+++ b/src/collectors/freebsd.plugin/freebsd_getmntinfo.c
diff --git a/collectors/freebsd.plugin/freebsd_ipfw.c b/src/collectors/freebsd.plugin/freebsd_ipfw.c
index dcb771ce9..dcb771ce9 100644
--- a/collectors/freebsd.plugin/freebsd_ipfw.c
+++ b/src/collectors/freebsd.plugin/freebsd_ipfw.c
diff --git a/collectors/freebsd.plugin/freebsd_kstat_zfs.c b/src/collectors/freebsd.plugin/freebsd_kstat_zfs.c
index 165efa17c..165efa17c 100644
--- a/collectors/freebsd.plugin/freebsd_kstat_zfs.c
+++ b/src/collectors/freebsd.plugin/freebsd_kstat_zfs.c
diff --git a/collectors/freebsd.plugin/freebsd_sysctl.c b/src/collectors/freebsd.plugin/freebsd_sysctl.c
index 8a6df509d..8a6df509d 100644
--- a/collectors/freebsd.plugin/freebsd_sysctl.c
+++ b/src/collectors/freebsd.plugin/freebsd_sysctl.c
diff --git a/collectors/freebsd.plugin/integrations/dev.cpu.0.freq.md b/src/collectors/freebsd.plugin/integrations/dev.cpu.0.freq.md
index 5f18661d0..59cfacb90 100644
--- a/collectors/freebsd.plugin/integrations/dev.cpu.0.freq.md
+++ b/src/collectors/freebsd.plugin/integrations/dev.cpu.0.freq.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/dev.cpu.0.freq.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/dev.cpu.0.freq.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "dev.cpu.0.freq"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -87,7 +87,7 @@ Configuration for this specific integration is located in the `[plugin:freebsd]`
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/dev.cpu.temperature.md b/src/collectors/freebsd.plugin/integrations/dev.cpu.temperature.md
index a3736f771..6e9c3d6dc 100644
--- a/collectors/freebsd.plugin/integrations/dev.cpu.temperature.md
+++ b/src/collectors/freebsd.plugin/integrations/dev.cpu.temperature.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/dev.cpu.temperature.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/dev.cpu.temperature.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "dev.cpu.temperature"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -96,7 +96,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/devstat.md b/src/collectors/freebsd.plugin/integrations/devstat.md
index 9d9c6400b..6c0a48cb4 100644
--- a/collectors/freebsd.plugin/integrations/devstat.md
+++ b/src/collectors/freebsd.plugin/integrations/devstat.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/devstat.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/devstat.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "devstat"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -93,7 +93,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 10min_disk_utilization ](https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf) | disk.util | average percentage of time ${label:device} disk was busy over the last 10 minutes |
+| [ 10min_disk_utilization ](https://github.com/netdata/netdata/blob/master/src/health/health.d/disks.conf) | disk.util | average percentage of time ${label:device} disk was busy over the last 10 minutes |
## Setup
@@ -120,7 +120,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/getifaddrs.md b/src/collectors/freebsd.plugin/integrations/getifaddrs.md
index 63c4ce136..3982c931b 100644
--- a/collectors/freebsd.plugin/integrations/getifaddrs.md
+++ b/src/collectors/freebsd.plugin/integrations/getifaddrs.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/getifaddrs.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/getifaddrs.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "getifaddrs"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -93,13 +93,13 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ interface_speed ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.net | network interface ${label:device} current speed |
-| [ inbound_packets_dropped_ratio ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.drops | ratio of inbound dropped packets for the network interface ${label:device} over the last 10 minutes |
-| [ outbound_packets_dropped_ratio ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.drops | ratio of outbound dropped packets for the network interface ${label:device} over the last 10 minutes |
-| [ 1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.packets | average number of packets received by the network interface ${label:device} over the last minute |
-| [ 10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
-| [ interface_inbound_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.errors | number of inbound errors for the network interface ${label:device} in the last 10 minutes |
-| [ interface_outbound_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.errors | number of outbound errors for the network interface ${label:device} in the last 10 minutes |
+| [ interface_speed ](https://github.com/netdata/netdata/blob/master/src/health/health.d/net.conf) | net.net | network interface ${label:device} current speed |
+| [ inbound_packets_dropped_ratio ](https://github.com/netdata/netdata/blob/master/src/health/health.d/net.conf) | net.drops | ratio of inbound dropped packets for the network interface ${label:device} over the last 10 minutes |
+| [ outbound_packets_dropped_ratio ](https://github.com/netdata/netdata/blob/master/src/health/health.d/net.conf) | net.drops | ratio of outbound dropped packets for the network interface ${label:device} over the last 10 minutes |
+| [ 1m_received_packets_rate ](https://github.com/netdata/netdata/blob/master/src/health/health.d/net.conf) | net.packets | average number of packets received by the network interface ${label:device} over the last minute |
+| [ 10s_received_packets_storm ](https://github.com/netdata/netdata/blob/master/src/health/health.d/net.conf) | net.packets | ratio of average number of received packets for the network interface ${label:device} over the last 10 seconds, compared to the rate over the last minute |
+| [ interface_inbound_errors ](https://github.com/netdata/netdata/blob/master/src/health/health.d/net.conf) | net.errors | number of inbound errors for the network interface ${label:device} in the last 10 minutes |
+| [ interface_outbound_errors ](https://github.com/netdata/netdata/blob/master/src/health/health.d/net.conf) | net.errors | number of outbound errors for the network interface ${label:device} in the last 10 minutes |
## Setup
@@ -126,7 +126,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/getmntinfo.md b/src/collectors/freebsd.plugin/integrations/getmntinfo.md
index d26ad1c03..7175dc74d 100644
--- a/collectors/freebsd.plugin/integrations/getmntinfo.md
+++ b/src/collectors/freebsd.plugin/integrations/getmntinfo.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/getmntinfo.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/getmntinfo.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "getmntinfo"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -75,8 +75,8 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ disk_space_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf) | disk.space | disk ${label:mount_point} space utilization |
-| [ disk_inode_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf) | disk.inodes | disk ${label:mount_point} inode utilization |
+| [ disk_space_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/disks.conf) | disk.space | disk ${label:mount_point} space utilization |
+| [ disk_inode_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/disks.conf) | disk.inodes | disk ${label:mount_point} inode utilization |
## Setup
@@ -103,7 +103,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/hw.intrcnt.md b/src/collectors/freebsd.plugin/integrations/hw.intrcnt.md
index 49164c369..5f39904bb 100644
--- a/collectors/freebsd.plugin/integrations/hw.intrcnt.md
+++ b/src/collectors/freebsd.plugin/integrations/hw.intrcnt.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/hw.intrcnt.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/hw.intrcnt.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "hw.intrcnt"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -97,7 +97,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/ipfw.md b/src/collectors/freebsd.plugin/integrations/ipfw.md
index 84e023bdf..ecbfbfbe3 100644
--- a/collectors/freebsd.plugin/integrations/ipfw.md
+++ b/src/collectors/freebsd.plugin/integrations/ipfw.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/ipfw.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/ipfw.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "ipfw"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -100,7 +100,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/kern.cp_time.md b/src/collectors/freebsd.plugin/integrations/kern.cp_time.md
index 95bdb8d90..b12ad8eda 100644
--- a/collectors/freebsd.plugin/integrations/kern.cp_time.md
+++ b/src/collectors/freebsd.plugin/integrations/kern.cp_time.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/kern.cp_time.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/kern.cp_time.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "kern.cp_time"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -86,10 +86,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf) | system.cpu | average CPU utilization over the last 10 minutes (excluding iowait, nice and steal) |
-| [ 10min_cpu_iowait ](https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf) | system.cpu | average CPU iowait time over the last 10 minutes |
-| [ 20min_steal_cpu ](https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf) | system.cpu | average CPU steal time over the last 20 minutes |
-| [ 10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf) | system.cpu | average CPU utilization over the last 10 minutes (excluding nice) |
+| [ 10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cpu.conf) | system.cpu | average CPU utilization over the last 10 minutes (excluding iowait, nice and steal) |
+| [ 10min_cpu_iowait ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cpu.conf) | system.cpu | average CPU iowait time over the last 10 minutes |
+| [ 20min_steal_cpu ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cpu.conf) | system.cpu | average CPU steal time over the last 20 minutes |
+| [ 10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cpu.conf) | system.cpu | average CPU utilization over the last 10 minutes (excluding nice) |
## Setup
@@ -115,7 +115,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/kern.ipc.msq.md b/src/collectors/freebsd.plugin/integrations/kern.ipc.msq.md
index e7457e0c1..6757495a2 100644
--- a/collectors/freebsd.plugin/integrations/kern.ipc.msq.md
+++ b/src/collectors/freebsd.plugin/integrations/kern.ipc.msq.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/kern.ipc.msq.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/kern.ipc.msq.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "kern.ipc.msq"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -98,7 +98,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/kern.ipc.sem.md b/src/collectors/freebsd.plugin/integrations/kern.ipc.sem.md
index 7bf7235e6..169c4a6bd 100644
--- a/collectors/freebsd.plugin/integrations/kern.ipc.sem.md
+++ b/src/collectors/freebsd.plugin/integrations/kern.ipc.sem.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/kern.ipc.sem.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/kern.ipc.sem.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "kern.ipc.sem"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -75,8 +75,8 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ semaphores_used ](https://github.com/netdata/netdata/blob/master/health/health.d/ipc.conf) | system.ipc_semaphores | IPC semaphore utilization |
-| [ semaphore_arrays_used ](https://github.com/netdata/netdata/blob/master/health/health.d/ipc.conf) | system.ipc_semaphore_arrays | IPC semaphore arrays utilization |
+| [ semaphores_used ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ipc.conf) | system.ipc_semaphores | IPC semaphore utilization |
+| [ semaphore_arrays_used ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ipc.conf) | system.ipc_semaphore_arrays | IPC semaphore arrays utilization |
## Setup
@@ -103,7 +103,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/kern.ipc.shm.md b/src/collectors/freebsd.plugin/integrations/kern.ipc.shm.md
index 1f10c1e6e..ac58b8228 100644
--- a/collectors/freebsd.plugin/integrations/kern.ipc.shm.md
+++ b/src/collectors/freebsd.plugin/integrations/kern.ipc.shm.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/kern.ipc.shm.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/kern.ipc.shm.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "kern.ipc.shm"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -97,7 +97,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/net.inet.icmp.stats.md b/src/collectors/freebsd.plugin/integrations/net.inet.icmp.stats.md
index 29562bc9a..e2bc6c998 100644
--- a/collectors/freebsd.plugin/integrations/net.inet.icmp.stats.md
+++ b/src/collectors/freebsd.plugin/integrations/net.inet.icmp.stats.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/net.inet.icmp.stats.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/net.inet.icmp.stats.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "net.inet.icmp.stats"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -98,7 +98,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/net.inet.ip.stats.md b/src/collectors/freebsd.plugin/integrations/net.inet.ip.stats.md
index 785767e89..299914ea6 100644
--- a/collectors/freebsd.plugin/integrations/net.inet.ip.stats.md
+++ b/src/collectors/freebsd.plugin/integrations/net.inet.ip.stats.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/net.inet.ip.stats.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/net.inet.ip.stats.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "net.inet.ip.stats"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -99,7 +99,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/net.inet.tcp.states.md b/src/collectors/freebsd.plugin/integrations/net.inet.tcp.states.md
index 5b4144580..2bb437156 100644
--- a/collectors/freebsd.plugin/integrations/net.inet.tcp.states.md
+++ b/src/collectors/freebsd.plugin/integrations/net.inet.tcp.states.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/net.inet.tcp.states.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/net.inet.tcp.states.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "net.inet.tcp.states"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -74,7 +74,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ tcp_connections ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_conn.conf) | ipv4.tcpsock | IPv4 TCP connections utilization |
+| [ tcp_connections ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_conn.conf) | ipv4.tcpsock | IPv4 TCP connections utilization |
## Setup
@@ -101,7 +101,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/net.inet.tcp.stats.md b/src/collectors/freebsd.plugin/integrations/net.inet.tcp.stats.md
index be779740d..6aa2a2239 100644
--- a/collectors/freebsd.plugin/integrations/net.inet.tcp.stats.md
+++ b/src/collectors/freebsd.plugin/integrations/net.inet.tcp.stats.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/net.inet.tcp.stats.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/net.inet.tcp.stats.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "net.inet.tcp.stats"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -81,10 +81,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 1m_ipv4_tcp_resets_sent ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf) | ipv4.tcphandshake | average number of sent TCP RESETS over the last minute |
-| [ 10s_ipv4_tcp_resets_sent ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf) | ipv4.tcphandshake | average number of sent TCP RESETS over the last 10 seconds. This can indicate a port scan, or that a service running on this host has crashed. Netdata will not send a clear notification for this alarm. |
-| [ 1m_ipv4_tcp_resets_received ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf) | ipv4.tcphandshake | average number of received TCP RESETS over the last minute |
-| [ 10s_ipv4_tcp_resets_received ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf) | ipv4.tcphandshake | average number of received TCP RESETS over the last 10 seconds. This can be an indication that a service this host needs has crashed. Netdata will not send a clear notification for this alarm. |
+| [ 1m_ipv4_tcp_resets_sent ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_resets.conf) | ipv4.tcphandshake | average number of sent TCP RESETS over the last minute |
+| [ 10s_ipv4_tcp_resets_sent ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_resets.conf) | ipv4.tcphandshake | average number of sent TCP RESETS over the last 10 seconds. This can indicate a port scan, or that a service running on this host has crashed. Netdata will not send a clear notification for this alarm. |
+| [ 1m_ipv4_tcp_resets_received ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_resets.conf) | ipv4.tcphandshake | average number of received TCP RESETS over the last minute |
+| [ 10s_ipv4_tcp_resets_received ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_resets.conf) | ipv4.tcphandshake | average number of received TCP RESETS over the last 10 seconds. This can be an indication that a service this host needs has crashed. Netdata will not send a clear notification for this alarm. |
## Setup
@@ -111,7 +111,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/net.inet.udp.stats.md b/src/collectors/freebsd.plugin/integrations/net.inet.udp.stats.md
index d3da40455..1745485e2 100644
--- a/collectors/freebsd.plugin/integrations/net.inet.udp.stats.md
+++ b/src/collectors/freebsd.plugin/integrations/net.inet.udp.stats.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/net.inet.udp.stats.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/net.inet.udp.stats.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "net.inet.udp.stats"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -75,8 +75,8 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 1m_ipv4_udp_receive_buffer_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/udp_errors.conf) | ipv4.udperrors | average number of UDP receive buffer errors over the last minute |
-| [ 1m_ipv4_udp_send_buffer_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/udp_errors.conf) | ipv4.udperrors | average number of UDP send buffer errors over the last minute |
+| [ 1m_ipv4_udp_receive_buffer_errors ](https://github.com/netdata/netdata/blob/master/src/health/health.d/udp_errors.conf) | ipv4.udperrors | average number of UDP receive buffer errors over the last minute |
+| [ 1m_ipv4_udp_send_buffer_errors ](https://github.com/netdata/netdata/blob/master/src/health/health.d/udp_errors.conf) | ipv4.udperrors | average number of UDP send buffer errors over the last minute |
## Setup
@@ -103,7 +103,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/net.inet6.icmp6.stats.md b/src/collectors/freebsd.plugin/integrations/net.inet6.icmp6.stats.md
index 7344b79b3..8d1578d30 100644
--- a/collectors/freebsd.plugin/integrations/net.inet6.icmp6.stats.md
+++ b/src/collectors/freebsd.plugin/integrations/net.inet6.icmp6.stats.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/net.inet6.icmp6.stats.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/net.inet6.icmp6.stats.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "net.inet6.icmp6.stats"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -102,7 +102,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/net.inet6.ip6.stats.md b/src/collectors/freebsd.plugin/integrations/net.inet6.ip6.stats.md
index d9128b529..d993cb5fb 100644
--- a/collectors/freebsd.plugin/integrations/net.inet6.ip6.stats.md
+++ b/src/collectors/freebsd.plugin/integrations/net.inet6.ip6.stats.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/net.inet6.ip6.stats.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/net.inet6.ip6.stats.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "net.inet6.ip6.stats"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -99,7 +99,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/net.isr.md b/src/collectors/freebsd.plugin/integrations/net.isr.md
index 2d75b825a..1ddfe64a1 100644
--- a/collectors/freebsd.plugin/integrations/net.isr.md
+++ b/src/collectors/freebsd.plugin/integrations/net.isr.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/net.isr.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/net.isr.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "net.isr"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -86,9 +86,9 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 1min_netdev_backlog_exceeded ](https://github.com/netdata/netdata/blob/master/health/health.d/softnet.conf) | system.softnet_stat | average number of dropped packets in the last minute due to exceeded net.core.netdev_max_backlog |
-| [ 1min_netdev_budget_ran_outs ](https://github.com/netdata/netdata/blob/master/health/health.d/softnet.conf) | system.softnet_stat | average number of times ksoftirq ran out of sysctl net.core.netdev_budget or net.core.netdev_budget_usecs with work remaining over the last minute (this can be a cause for dropped packets) |
-| [ 10min_netisr_backlog_exceeded ](https://github.com/netdata/netdata/blob/master/health/health.d/softnet.conf) | system.softnet_stat | average number of drops in the last minute due to exceeded sysctl net.route.netisr_maxqlen (this can be a cause for dropped packets) |
+| [ 1min_netdev_backlog_exceeded ](https://github.com/netdata/netdata/blob/master/src/health/health.d/softnet.conf) | system.softnet_stat | average number of dropped packets in the last minute due to exceeded net.core.netdev_max_backlog |
+| [ 1min_netdev_budget_ran_outs ](https://github.com/netdata/netdata/blob/master/src/health/health.d/softnet.conf) | system.softnet_stat | average number of times ksoftirq ran out of sysctl net.core.netdev_budget or net.core.netdev_budget_usecs with work remaining over the last minute (this can be a cause for dropped packets) |
+| [ 10min_netisr_backlog_exceeded ](https://github.com/netdata/netdata/blob/master/src/health/health.d/softnet.conf) | system.softnet_stat | average number of drops in the last minute due to exceeded sysctl net.route.netisr_maxqlen (this can be a cause for dropped packets) |
## Setup
@@ -115,7 +115,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/system.ram.md b/src/collectors/freebsd.plugin/integrations/system.ram.md
index 7d4974922..8f7759384 100644
--- a/collectors/freebsd.plugin/integrations/system.ram.md
+++ b/src/collectors/freebsd.plugin/integrations/system.ram.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/system.ram.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/system.ram.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "system.ram"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -75,10 +75,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ ram_in_use ](https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf) | system.ram | system memory utilization |
-| [ ram_in_use ](https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf) | system.ram | system memory utilization |
-| [ ram_available ](https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf) | mem.available | percentage of estimated amount of RAM available for userspace processes, without causing swapping |
-| [ ram_available ](https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf) | mem.available | percentage of estimated amount of RAM available for userspace processes, without causing swapping |
+| [ ram_in_use ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ram.conf) | system.ram | system memory utilization |
+| [ ram_in_use ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ram.conf) | system.ram | system memory utilization |
+| [ ram_available ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ram.conf) | mem.available | percentage of estimated amount of RAM available for userspace processes, without causing swapping |
+| [ ram_available ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ram.conf) | mem.available | percentage of estimated amount of RAM available for userspace processes, without causing swapping |
## Setup
@@ -105,7 +105,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/uptime.md b/src/collectors/freebsd.plugin/integrations/uptime.md
index e3f1db3f1..285cd7387 100644
--- a/collectors/freebsd.plugin/integrations/uptime.md
+++ b/src/collectors/freebsd.plugin/integrations/uptime.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/uptime.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/uptime.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "uptime"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -96,7 +96,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/vm.loadavg.md b/src/collectors/freebsd.plugin/integrations/vm.loadavg.md
index 88c47b7a4..25091785d 100644
--- a/collectors/freebsd.plugin/integrations/vm.loadavg.md
+++ b/src/collectors/freebsd.plugin/integrations/vm.loadavg.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/vm.loadavg.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/vm.loadavg.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "vm.loadavg"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -74,10 +74,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ load_cpu_number ](https://github.com/netdata/netdata/blob/master/health/health.d/load.conf) | system.load | number of active CPU cores in the system |
-| [ load_average_15 ](https://github.com/netdata/netdata/blob/master/health/health.d/load.conf) | system.load | system fifteen-minute load average |
-| [ load_average_5 ](https://github.com/netdata/netdata/blob/master/health/health.d/load.conf) | system.load | system five-minute load average |
-| [ load_average_1 ](https://github.com/netdata/netdata/blob/master/health/health.d/load.conf) | system.load | system one-minute load average |
+| [ load_cpu_number ](https://github.com/netdata/netdata/blob/master/src/health/health.d/load.conf) | system.load | number of active CPU cores in the system |
+| [ load_average_15 ](https://github.com/netdata/netdata/blob/master/src/health/health.d/load.conf) | system.load | system fifteen-minute load average |
+| [ load_average_5 ](https://github.com/netdata/netdata/blob/master/src/health/health.d/load.conf) | system.load | system five-minute load average |
+| [ load_average_1 ](https://github.com/netdata/netdata/blob/master/src/health/health.d/load.conf) | system.load | system one-minute load average |
## Setup
@@ -104,7 +104,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/vm.stats.sys.v_intr.md b/src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_intr.md
index c3e7466e9..fdf2a8be0 100644
--- a/collectors/freebsd.plugin/integrations/vm.stats.sys.v_intr.md
+++ b/src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_intr.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/vm.stats.sys.v_intr.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_intr.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "vm.stats.sys.v_intr"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -96,7 +96,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/vm.stats.sys.v_soft.md b/src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_soft.md
index ce914bb50..679cd974c 100644
--- a/collectors/freebsd.plugin/integrations/vm.stats.sys.v_soft.md
+++ b/src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_soft.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/vm.stats.sys.v_soft.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_soft.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "vm.stats.sys.v_soft"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -96,7 +96,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/vm.stats.sys.v_swtch.md b/src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_swtch.md
index cbcee311f..51ad841fb 100644
--- a/collectors/freebsd.plugin/integrations/vm.stats.sys.v_swtch.md
+++ b/src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_swtch.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/vm.stats.sys.v_swtch.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/vm.stats.sys.v_swtch.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "vm.stats.sys.v_swtch"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -97,7 +97,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/vm.stats.vm.v_pgfaults.md b/src/collectors/freebsd.plugin/integrations/vm.stats.vm.v_pgfaults.md
index 19230dd56..313724c70 100644
--- a/collectors/freebsd.plugin/integrations/vm.stats.vm.v_pgfaults.md
+++ b/src/collectors/freebsd.plugin/integrations/vm.stats.vm.v_pgfaults.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/vm.stats.vm.v_pgfaults.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/vm.stats.vm.v_pgfaults.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "vm.stats.vm.v_pgfaults"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -96,7 +96,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/vm.stats.vm.v_swappgs.md b/src/collectors/freebsd.plugin/integrations/vm.stats.vm.v_swappgs.md
index c6caaa682..76ae46463 100644
--- a/collectors/freebsd.plugin/integrations/vm.stats.vm.v_swappgs.md
+++ b/src/collectors/freebsd.plugin/integrations/vm.stats.vm.v_swappgs.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/vm.stats.vm.v_swappgs.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/vm.stats.vm.v_swappgs.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "vm.stats.vm.v_swappgs"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -74,7 +74,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 30min_ram_swapped_out ](https://github.com/netdata/netdata/blob/master/health/health.d/swap.conf) | mem.swapio | percentage of the system RAM swapped in the last 30 minutes |
+| [ 30min_ram_swapped_out ](https://github.com/netdata/netdata/blob/master/src/health/health.d/swap.conf) | mem.swapio | percentage of the system RAM swapped in the last 30 minutes |
## Setup
@@ -101,7 +101,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/vm.swap_info.md b/src/collectors/freebsd.plugin/integrations/vm.swap_info.md
index caa22b3dc..e79554818 100644
--- a/collectors/freebsd.plugin/integrations/vm.swap_info.md
+++ b/src/collectors/freebsd.plugin/integrations/vm.swap_info.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/vm.swap_info.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/vm.swap_info.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "vm.swap_info"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -74,7 +74,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ used_swap ](https://github.com/netdata/netdata/blob/master/health/health.d/swap.conf) | mem.swap | swap memory utilization |
+| [ used_swap ](https://github.com/netdata/netdata/blob/master/src/health/health.d/swap.conf) | mem.swap | swap memory utilization |
## Setup
@@ -101,7 +101,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/vm.vmtotal.md b/src/collectors/freebsd.plugin/integrations/vm.vmtotal.md
index f3f631af6..322ce809e 100644
--- a/collectors/freebsd.plugin/integrations/vm.vmtotal.md
+++ b/src/collectors/freebsd.plugin/integrations/vm.vmtotal.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/vm.vmtotal.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/vm.vmtotal.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "vm.vmtotal"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -76,7 +76,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ active_processes ](https://github.com/netdata/netdata/blob/master/health/health.d/processes.conf) | system.active_processes | system process IDs (PID) space utilization |
+| [ active_processes ](https://github.com/netdata/netdata/blob/master/src/health/health.d/processes.conf) | system.active_processes | system process IDs (PID) space utilization |
## Setup
@@ -103,7 +103,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/integrations/zfs.md b/src/collectors/freebsd.plugin/integrations/zfs.md
index 99f10026d..88a8249e2 100644
--- a/collectors/freebsd.plugin/integrations/zfs.md
+++ b/src/collectors/freebsd.plugin/integrations/zfs.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/integrations/zfs.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freebsd.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/integrations/zfs.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freebsd.plugin/metadata.yaml"
sidebar_label: "zfs"
learn_status: "Published"
-learn_rel_path: "Data Collection/FreeBSD"
+learn_rel_path: "Collecting Metrics/FreeBSD"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -101,7 +101,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ zfs_memory_throttle ](https://github.com/netdata/netdata/blob/master/health/health.d/zfs.conf) | zfs.memory_ops | number of times ZFS had to limit the ARC growth in the last 10 minutes |
+| [ zfs_memory_throttle ](https://github.com/netdata/netdata/blob/master/src/health/health.d/zfs.conf) | zfs.memory_ops | number of times ZFS had to limit the ARC growth in the last 10 minutes |
## Setup
@@ -128,7 +128,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/freebsd.plugin/plugin_freebsd.c b/src/collectors/freebsd.plugin/plugin_freebsd.c
index 976fe26fb..6b77dcf3a 100644
--- a/collectors/freebsd.plugin/plugin_freebsd.c
+++ b/src/collectors/freebsd.plugin/plugin_freebsd.c
@@ -91,7 +91,7 @@ void *freebsd_main(void *ptr)
// initialize FreeBSD plugin
if (freebsd_plugin_init())
- netdata_cleanup_and_exit(1);
+ netdata_cleanup_and_exit(1, NULL, NULL, NULL);
// check the enabled status for each module
int i;
diff --git a/collectors/freebsd.plugin/plugin_freebsd.h b/src/collectors/freebsd.plugin/plugin_freebsd.h
index af7d0822e..af7d0822e 100644
--- a/collectors/freebsd.plugin/plugin_freebsd.h
+++ b/src/collectors/freebsd.plugin/plugin_freebsd.h
diff --git a/collectors/freeipmi.plugin/README.md b/src/collectors/freeipmi.plugin/README.md
index f55ebf73d..f55ebf73d 120000
--- a/collectors/freeipmi.plugin/README.md
+++ b/src/collectors/freeipmi.plugin/README.md
diff --git a/collectors/freeipmi.plugin/freeipmi_plugin.c b/src/collectors/freeipmi.plugin/freeipmi_plugin.c
index 6ec9b698b..3b27c5d5a 100644
--- a/collectors/freeipmi.plugin/freeipmi_plugin.c
+++ b/src/collectors/freeipmi.plugin/freeipmi_plugin.c
@@ -23,7 +23,9 @@
#include "libnetdata/required_dummies.h"
#define FREEIPMI_GLOBAL_FUNCTION_SENSORS() do { \
- fprintf(stdout, PLUGINSD_KEYWORD_FUNCTION " GLOBAL \"ipmi-sensors\" %d \"%s\"\n", 5, "Displays current sensor state and readings"); \
+ fprintf(stdout, PLUGINSD_KEYWORD_FUNCTION " GLOBAL \"ipmi-sensors\" %d \"%s\" \"top\" "HTTP_ACCESS_FORMAT" %d\n", \
+ 5, "Displays current sensor state and readings", \
+ (HTTP_ACCESS_FORMAT_CAST)(HTTP_ACCESS_NONE), 100); \
} while(0)
// component names, based on our patterns
@@ -65,9 +67,8 @@ static void netdata_update_ipmi_sensor_reading(
, int sensor_bitmask_type
, int sensor_bitmask
, char **sensor_bitmask_strings
- , struct netdata_ipmi_state *state
-);
-static void netdata_update_ipmi_sel_events_count(struct netdata_ipmi_state *state, uint32_t events);
+ , struct netdata_ipmi_state *stt);
+static void netdata_update_ipmi_sel_events_count(struct netdata_ipmi_state *stt, uint32_t events);
// END NETDATA CODE
// ----------------------------------------------------------------------------
@@ -905,7 +906,7 @@ const char *netdata_collect_type_to_string(IPMI_COLLECTION_TYPE type) {
return "unknown";
}
-static void netdata_sensor_set_value(struct sensor *sn, void *sensor_reading, struct netdata_ipmi_state *state __maybe_unused) {
+static void netdata_sensor_set_value(struct sensor *sn, void *sensor_reading, struct netdata_ipmi_state *stt __maybe_unused) {
switch(sn->sensor_reading_type) {
case IPMI_MONITORING_SENSOR_READING_TYPE_UNSIGNED_INTEGER8_BOOL:
sn->sensor_reading.bool_value = *((uint8_t *)sensor_reading);
@@ -939,8 +940,7 @@ static void netdata_update_ipmi_sensor_reading(
, int sensor_bitmask_type __maybe_unused
, int sensor_bitmask __maybe_unused
, char **sensor_bitmask_strings __maybe_unused
- , struct netdata_ipmi_state *state
-) {
+ , struct netdata_ipmi_state *stt) {
if(unlikely(sensor_state == IPMI_MONITORING_STATE_UNKNOWN &&
sensor_type == IPMI_MONITORING_SENSOR_TYPE_UNKNOWN &&
sensor_units == IPMI_MONITORING_SENSOR_UNITS_UNKNOWN &&
@@ -952,38 +952,38 @@ static void netdata_update_ipmi_sensor_reading(
if(unlikely(!sensor_name || !*sensor_name))
sensor_name = "UNNAMED";
- state->sensors.collected++;
+ stt->sensors.collected++;
char key[SENSORS_DICT_KEY_SIZE + 1];
snprintfz(key, SENSORS_DICT_KEY_SIZE, "i%d_n%d_t%d_u%d_%s",
record_id, sensor_number, sensor_reading_type, sensor_units, sensor_name);
// find the sensor record
- const DICTIONARY_ITEM *item = dictionary_get_and_acquire_item(state->sensors.dict, key);
+ const DICTIONARY_ITEM *item = dictionary_get_and_acquire_item(stt->sensors.dict, key);
if(likely(item)) {
// recurring collection
- if(state->debug)
+ if(stt->debug)
fprintf(stderr, "%s: reusing sensor record for sensor '%s', id %d, number %d, type %d, state %d, units %d, reading_type %d\n",
program_name, sensor_name, record_id, sensor_number, sensor_type, sensor_state, sensor_units, sensor_reading_type);
struct sensor *sn = dictionary_acquired_item_value(item);
if(sensor_reading) {
- netdata_sensor_set_value(sn, sensor_reading, state);
- sn->last_collected_metric_ut = state->sensors.now_ut;
+ netdata_sensor_set_value(sn, sensor_reading, stt);
+ sn->last_collected_metric_ut = stt->sensors.now_ut;
}
sn->sensor_state = sensor_state;
- sn->last_collected_state_ut = state->sensors.now_ut;
+ sn->last_collected_state_ut = stt->sensors.now_ut;
- dictionary_acquired_item_release(state->sensors.dict, item);
+ dictionary_acquired_item_release(stt->sensors.dict, item);
return;
}
- if(state->debug)
+ if(stt->debug)
fprintf(stderr, "Allocating new sensor data record for sensor '%s', id %d, number %d, type %d, state %d, units %d, reading_type %d\n",
sensor_name, record_id, sensor_number, sensor_type, sensor_state, sensor_units, sensor_reading_type);
@@ -992,12 +992,12 @@ static void netdata_update_ipmi_sensor_reading(
bool excluded_state = excluded_status_record_ids_check(record_id);
if(excluded_metric) {
- if(state->debug)
+ if(stt->debug)
fprintf(stderr, "Sensor '%s' is excluded by excluded_record_ids_check()\n", sensor_name);
}
if(excluded_state) {
- if(state->debug)
+ if(stt->debug)
fprintf(stderr, "Sensor '%s' is excluded for status check, by excluded_status_record_ids_check()\n", sensor_name);
}
@@ -1022,7 +1022,7 @@ static void netdata_update_ipmi_sensor_reading(
t.units = "Celsius";
t.family = "temperatures";
t.chart_type = "line";
- t.priority = state->sensors.priority + 10;
+ t.priority = stt->sensors.priority + 10;
break;
case IPMI_MONITORING_SENSOR_UNITS_FAHRENHEIT:
@@ -1032,7 +1032,7 @@ static void netdata_update_ipmi_sensor_reading(
t.units = "Fahrenheit";
t.family = "temperatures";
t.chart_type = "line";
- t.priority = state->sensors.priority + 20;
+ t.priority = stt->sensors.priority + 20;
break;
case IPMI_MONITORING_SENSOR_UNITS_VOLTS:
@@ -1042,7 +1042,7 @@ static void netdata_update_ipmi_sensor_reading(
t.units = "Volts";
t.family = "voltages";
t.chart_type = "line";
- t.priority = state->sensors.priority + 30;
+ t.priority = stt->sensors.priority + 30;
break;
case IPMI_MONITORING_SENSOR_UNITS_AMPS:
@@ -1052,7 +1052,7 @@ static void netdata_update_ipmi_sensor_reading(
t.units = "Amps";
t.family = "current";
t.chart_type = "line";
- t.priority = state->sensors.priority + 40;
+ t.priority = stt->sensors.priority + 40;
break;
case IPMI_MONITORING_SENSOR_UNITS_RPM:
@@ -1062,7 +1062,7 @@ static void netdata_update_ipmi_sensor_reading(
t.units = "RPM";
t.family = "fans";
t.chart_type = "line";
- t.priority = state->sensors.priority + 50;
+ t.priority = stt->sensors.priority + 50;
break;
case IPMI_MONITORING_SENSOR_UNITS_WATTS:
@@ -1072,7 +1072,7 @@ static void netdata_update_ipmi_sensor_reading(
t.units = "Watts";
t.family = "power";
t.chart_type = "line";
- t.priority = state->sensors.priority + 60;
+ t.priority = stt->sensors.priority + 60;
break;
case IPMI_MONITORING_SENSOR_UNITS_PERCENT:
@@ -1082,11 +1082,11 @@ static void netdata_update_ipmi_sensor_reading(
t.units = "%%";
t.family = "other";
t.chart_type = "line";
- t.priority = state->sensors.priority + 70;
+ t.priority = stt->sensors.priority + 70;
break;
default:
- t.priority = state->sensors.priority + 80;
+ t.priority = stt->sensors.priority + 80;
t.do_metric = false;
break;
}
@@ -1107,57 +1107,57 @@ static void netdata_update_ipmi_sensor_reading(
}
if(sensor_reading) {
- netdata_sensor_set_value(&t, sensor_reading, state);
- t.last_collected_metric_ut = state->sensors.now_ut;
+ netdata_sensor_set_value(&t, sensor_reading, stt);
+ t.last_collected_metric_ut = stt->sensors.now_ut;
}
- t.last_collected_state_ut = state->sensors.now_ut;
+ t.last_collected_state_ut = stt->sensors.now_ut;
- dictionary_set(state->sensors.dict, key, &t, sizeof(t));
+ dictionary_set(stt->sensors.dict, key, &t, sizeof(t));
}
-static void netdata_update_ipmi_sel_events_count(struct netdata_ipmi_state *state, uint32_t events) {
- state->sel.events = events;
+static void netdata_update_ipmi_sel_events_count(struct netdata_ipmi_state *stt, uint32_t events) {
+ stt->sel.events = events;
}
-int netdata_ipmi_collect_data(struct ipmi_monitoring_ipmi_config *ipmi_config, IPMI_COLLECTION_TYPE type, struct netdata_ipmi_state *state) {
+int netdata_ipmi_collect_data(struct ipmi_monitoring_ipmi_config *ipmi_config, IPMI_COLLECTION_TYPE type, struct netdata_ipmi_state *stt) {
errno = 0;
if(type & IPMI_COLLECT_TYPE_SENSORS) {
- state->sensors.collected = 0;
- state->sensors.now_ut = now_monotonic_usec();
+ stt->sensors.collected = 0;
+ stt->sensors.now_ut = now_monotonic_usec();
- if (netdata_read_ipmi_sensors(ipmi_config, state) < 0) return -1;
+ if (netdata_read_ipmi_sensors(ipmi_config, stt) < 0) return -1;
}
if(type & IPMI_COLLECT_TYPE_SEL) {
- state->sel.events = 0;
- state->sel.now_ut = now_monotonic_usec();
- if(netdata_get_ipmi_sel_events_count(ipmi_config, state) < 0) return -2;
+ stt->sel.events = 0;
+ stt->sel.now_ut = now_monotonic_usec();
+ if(netdata_get_ipmi_sel_events_count(ipmi_config, stt) < 0) return -2;
}
return 0;
}
-int netdata_ipmi_detect_speed_secs(struct ipmi_monitoring_ipmi_config *ipmi_config, IPMI_COLLECTION_TYPE type, struct netdata_ipmi_state *state) {
+int netdata_ipmi_detect_speed_secs(struct ipmi_monitoring_ipmi_config *ipmi_config, IPMI_COLLECTION_TYPE type, struct netdata_ipmi_state *stt) {
int i, checks = SPEED_TEST_ITERATIONS, successful = 0;
usec_t total = 0;
for(i = 0 ; i < checks ; i++) {
- if(unlikely(state->debug))
+ if(unlikely(stt->debug))
fprintf(stderr, "%s: checking %s data collection speed iteration %d of %d\n",
program_name, netdata_collect_type_to_string(type), i + 1, checks);
// measure the time a data collection needs
usec_t start = now_realtime_usec();
- if(netdata_ipmi_collect_data(ipmi_config, type, state) < 0)
+ if(netdata_ipmi_collect_data(ipmi_config, type, stt) < 0)
continue;
usec_t end = now_realtime_usec();
successful++;
- if(unlikely(state->debug))
+ if(unlikely(stt->debug))
fprintf(stderr, "%s: %s data collection speed was %"PRIu64" usec\n",
program_name, netdata_collect_type_to_string(type), end - start);
@@ -1297,31 +1297,32 @@ static inline bool is_sensor_updated(usec_t last_collected_ut, usec_t now_ut, us
return (now_ut - last_collected_ut < freq * 2) ? true : false;
}
-static size_t send_ipmi_sensor_metrics_to_netdata(struct netdata_ipmi_state *state) {
- if(state->sensors.status != ICS_RUNNING) {
- if(unlikely(state->debug))
+static size_t send_ipmi_sensor_metrics_to_netdata(struct netdata_ipmi_state *stt) {
+ if(stt->sensors.status != ICS_RUNNING) {
+ if(unlikely(stt->debug))
fprintf(stderr, "%s: %s() sensors state is not RUNNING\n",
program_name, __FUNCTION__ );
return 0;
}
size_t total_sensors_sent = 0;
- int update_every = (int)(state->sensors.freq_ut / USEC_PER_SEC);
+ int update_every_s = (int)(stt->sensors.freq_ut / USEC_PER_SEC);
struct sensor *sn;
netdata_mutex_lock(&stdout_mutex);
// generate the CHART/DIMENSION lines, if we have to
- dfe_start_reentrant(state->sensors.dict, sn) {
+ dfe_start_reentrant(stt->sensors.dict, sn) {
if(unlikely(!sn->do_metric && !sn->do_state))
continue;
bool did_metric = false, did_state = false;
if(likely(sn->do_metric)) {
- if(unlikely(!is_sensor_updated(sn->last_collected_metric_ut, state->updates.now_ut, state->sensors.freq_ut))) {
- if(unlikely(state->debug))
+ if(unlikely(!is_sensor_updated(sn->last_collected_metric_ut, stt->updates.now_ut, stt->sensors.freq_ut))) {
+ if(unlikely(stt->debug))
fprintf(stderr, "%s: %s() sensor '%s' metric is not UPDATED (last updated %"PRIu64", now %"PRIu64", freq %"PRIu64"\n",
- program_name, __FUNCTION__, sn->sensor_name, sn->last_collected_metric_ut, state->updates.now_ut, state->sensors.freq_ut);
+ program_name, __FUNCTION__, sn->sensor_name, sn->last_collected_metric_ut,
+ stt->updates.now_ut, stt->sensors.freq_ut);
}
else {
if (unlikely(!sn->metric_chart_sent)) {
@@ -1329,7 +1330,8 @@ static size_t send_ipmi_sensor_metrics_to_netdata(struct netdata_ipmi_state *sta
printf("CHART '%s_%s' '' '%s' '%s' '%s' '%s' '%s' %d %d '' '%s' '%s'\n",
sn->context, sn_dfe.name, sn->title, sn->units, sn->family, sn->context,
- sn->chart_type, sn->priority + 1, update_every, program_name, "sensors");
+ sn->chart_type, sn->priority + 1,
+ update_every_s, program_name, "sensors");
printf("CLABEL 'sensor' '%s' 1\n", sn->sensor_name);
printf("CLABEL 'type' '%s' 1\n", sn->type);
@@ -1343,19 +1345,16 @@ static size_t send_ipmi_sensor_metrics_to_netdata(struct netdata_ipmi_state *sta
switch (sn->sensor_reading_type) {
case IPMI_MONITORING_SENSOR_READING_TYPE_UNSIGNED_INTEGER32:
- printf("SET '%s' = %u\n", sn->dimension, sn->sensor_reading.uint32_value
- );
+ printf("SET '%s' = %u\n", sn->dimension, sn->sensor_reading.uint32_value);
break;
case IPMI_MONITORING_SENSOR_READING_TYPE_DOUBLE:
printf("SET '%s' = %lld\n", sn->dimension,
- (long long int) (sn->sensor_reading.double_value * sn->multiplier)
- );
+ (long long int) (sn->sensor_reading.double_value * sn->multiplier));
break;
case IPMI_MONITORING_SENSOR_READING_TYPE_UNSIGNED_INTEGER8_BOOL:
- printf("SET '%s' = %u\n", sn->dimension, sn->sensor_reading.bool_value
- );
+ printf("SET '%s' = %u\n", sn->dimension, sn->sensor_reading.bool_value);
break;
default:
@@ -1371,17 +1370,18 @@ static size_t send_ipmi_sensor_metrics_to_netdata(struct netdata_ipmi_state *sta
}
if(likely(sn->do_state)) {
- if(unlikely(!is_sensor_updated(sn->last_collected_state_ut, state->updates.now_ut, state->sensors.freq_ut))) {
- if (unlikely(state->debug))
+ if(unlikely(!is_sensor_updated(sn->last_collected_state_ut, stt->updates.now_ut, stt->sensors.freq_ut))) {
+ if (unlikely(stt->debug))
fprintf(stderr, "%s: %s() sensor '%s' state is not UPDATED (last updated %"PRIu64", now %"PRIu64", freq %"PRIu64"\n",
- program_name, __FUNCTION__, sn->sensor_name, sn->last_collected_state_ut, state->updates.now_ut, state->sensors.freq_ut);
+ program_name, __FUNCTION__, sn->sensor_name, sn->last_collected_state_ut,
+ stt->updates.now_ut, stt->sensors.freq_ut);
}
else {
if (unlikely(!sn->state_chart_sent)) {
sn->state_chart_sent = true;
printf("CHART 'ipmi.sensor_state_%s' '' 'IPMI Sensor State' 'state' 'states' 'ipmi.sensor_state' 'line' %d %d '' '%s' '%s'\n",
- sn_dfe.name, sn->priority, update_every, program_name, "sensors");
+ sn_dfe.name, sn->priority, update_every_s, program_name, "sensors");
printf("CLABEL 'sensor' '%s' 1\n", sn->sensor_name);
printf("CLABEL 'type' '%s' 1\n", sn->type);
@@ -1414,17 +1414,17 @@ static size_t send_ipmi_sensor_metrics_to_netdata(struct netdata_ipmi_state *sta
return total_sensors_sent;
}
-static size_t send_ipmi_sel_metrics_to_netdata(struct netdata_ipmi_state *state) {
+static size_t send_ipmi_sel_metrics_to_netdata(struct netdata_ipmi_state *stt) {
static bool sel_chart_generated = false;
netdata_mutex_lock(&stdout_mutex);
- if(likely(state->sel.status == ICS_RUNNING)) {
+ if(likely(stt->sel.status == ICS_RUNNING)) {
if(unlikely(!sel_chart_generated)) {
sel_chart_generated = true;
printf("CHART ipmi.events '' 'IPMI Events' 'events' 'events' ipmi.sel area %d %d '' '%s' '%s'\n"
- , state->sel.priority + 2
- , (int)(state->sel.freq_ut / USEC_PER_SEC)
+ , stt->sel.priority + 2
+ , (int)(stt->sel.freq_ut / USEC_PER_SEC)
, program_name
, "sel"
);
@@ -1435,13 +1435,14 @@ static size_t send_ipmi_sel_metrics_to_netdata(struct netdata_ipmi_state *state)
"BEGIN ipmi.events\n"
"SET events = %zu\n"
"END\n"
- , state->sel.events
+ ,
+ stt->sel.events
);
}
netdata_mutex_unlock(&stdout_mutex);
- return state->sel.events;
+ return stt->sel.events;
}
// ----------------------------------------------------------------------------
@@ -1470,15 +1471,35 @@ static const char *get_sensor_function_priority(struct sensor *sn) {
}
}
-static void freeimi_function_sensors(const char *transaction, char *function __maybe_unused, int timeout __maybe_unused, bool *cancelled __maybe_unused) {
- time_t expires = now_realtime_sec() + update_every;
+static void freeimi_function_sensors(const char *transaction, char *function __maybe_unused,
+ usec_t *stop_monotonic_ut __maybe_unused, bool *cancelled __maybe_unused,
+ BUFFER *payload __maybe_unused, HTTP_ACCESS access __maybe_unused,
+ const char *source __maybe_unused, void *data __maybe_unused) {
+ time_t now_s = now_realtime_sec();
- BUFFER *wb = buffer_create(PLUGINSD_LINE_MAX, NULL);
+ BUFFER *wb = buffer_create(4096, NULL);
buffer_json_initialize(wb, "\"", "\"", 0, true, BUFFER_JSON_OPTIONS_NEWLINE_ON_ARRAY_ITEMS);
buffer_json_member_add_uint64(wb, "status", HTTP_RESP_OK);
buffer_json_member_add_string(wb, "type", "table");
buffer_json_member_add_time_t(wb, "update_every", update_every);
+ buffer_json_member_add_boolean(wb, "has_history", false);
buffer_json_member_add_string(wb, "help", "View IPMI sensor readings and its state");
+
+ char function_copy[strlen(function) + 1];
+ memcpy(function_copy, function, sizeof(function_copy));
+ char *words[1024];
+ size_t num_words = quoted_strings_splitter_pluginsd(function_copy, words, 1024);
+ for(size_t i = 1; i < num_words ;i++) {
+ char *param = get_word(words, num_words, i);
+ if(strcmp(param, "info") == 0) {
+ buffer_json_member_add_array(wb, "accepted_params");
+ buffer_json_array_close(wb); // accepted_params
+ buffer_json_member_add_array(wb, "required_params");
+ buffer_json_array_close(wb); // required_params
+ goto close_and_send;
+ }
+ }
+
buffer_json_member_add_array(wb, "data");
struct sensor *sn;
@@ -1604,10 +1625,11 @@ static void freeimi_function_sensors(const char *transaction, char *function __m
}
buffer_json_array_close(wb);
- buffer_json_member_add_time_t(wb, "expires", now_realtime_sec() + 1);
+close_and_send:
+ buffer_json_member_add_time_t(wb, "expires", now_s + update_every);
buffer_json_finalize(wb);
- pluginsd_function_result_to_stdout(transaction, HTTP_RESP_OK, "application/json", expires, wb);
+ pluginsd_function_result_to_stdout(transaction, HTTP_RESP_OK, "application/json", now_s + update_every, wb);
buffer_free(wb);
}
@@ -1615,7 +1637,7 @@ static void freeimi_function_sensors(const char *transaction, char *function __m
// ----------------------------------------------------------------------------
// main, command line arguments parsing
-static void plugin_exit(int code) {
+static NORETURN void plugin_exit(int code) {
fflush(stdout);
function_plugin_should_exit = true;
exit(code);
@@ -1803,7 +1825,7 @@ int main (int argc, char **argv) {
" options ipmi_si kipmid_max_busy_us=10\n"
"\n"
" For more information:\n"
- " https://github.com/netdata/netdata/tree/master/collectors/freeipmi.plugin\n"
+ " https://github.com/netdata/netdata/tree/master/src/collectors/freeipmi.plugin\n"
"\n"
, program_name, VERSION
, update_every
@@ -1972,7 +1994,7 @@ int main (int argc, char **argv) {
size_t iteration = 0;
usec_t step = 100 * USEC_PER_MS;
bool global_chart_created = false;
- bool tty = isatty(fileno(stderr)) == 1;
+ bool tty = isatty(fileno(stdout)) == 1;
heartbeat_t hb;
heartbeat_init(&hb);
@@ -2044,7 +2066,7 @@ int main (int argc, char **argv) {
struct functions_evloop_globals *wg =
functions_evloop_init(1, "FREEIPMI", &stdout_mutex, &function_plugin_should_exit);
functions_evloop_add_function(
- wg, "ipmi-sensors", freeimi_function_sensors, PLUGINS_FUNCTIONS_TIMEOUT_DEFAULT);
+ wg, "ipmi-sensors", freeimi_function_sensors, PLUGINS_FUNCTIONS_TIMEOUT_DEFAULT, NULL);
FREEIPMI_GLOBAL_FUNCTION_SENSORS();
}
diff --git a/collectors/freeipmi.plugin/integrations/intelligent_platform_management_interface_ipmi.md b/src/collectors/freeipmi.plugin/integrations/intelligent_platform_management_interface_ipmi.md
index c0293fc37..ad0f17029 100644
--- a/collectors/freeipmi.plugin/integrations/intelligent_platform_management_interface_ipmi.md
+++ b/src/collectors/freeipmi.plugin/integrations/intelligent_platform_management_interface_ipmi.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/freeipmi.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/freeipmi.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/freeipmi.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/freeipmi.plugin/metadata.yaml"
sidebar_label: "Intelligent Platform Management Interface (IPMI)"
learn_status: "Published"
-learn_rel_path: "Data Collection/Hardware Devices and Sensors"
+learn_rel_path: "Collecting Metrics/Hardware Devices and Sensors"
most_popular: True
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -103,7 +103,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ ipmi_sensor_state ](https://github.com/netdata/netdata/blob/master/health/health.d/ipmi.conf) | ipmi.sensor_state | IPMI sensor ${label:sensor} (${label:component}) state |
+| [ ipmi_sensor_state ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ipmi.conf) | ipmi.sensor_state | IPMI sensor ${label:sensor} (${label:component}) state |
## Setup
@@ -144,7 +144,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/idlejitter.plugin/README.md b/src/collectors/idlejitter.plugin/README.md
index 1ce460b62..1ce460b62 120000
--- a/collectors/idlejitter.plugin/README.md
+++ b/src/collectors/idlejitter.plugin/README.md
diff --git a/collectors/idlejitter.plugin/integrations/idle_os_jitter.md b/src/collectors/idlejitter.plugin/integrations/idle_os_jitter.md
index 44463f6f5..23a15d001 100644
--- a/collectors/idlejitter.plugin/integrations/idle_os_jitter.md
+++ b/src/collectors/idlejitter.plugin/integrations/idle_os_jitter.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/idlejitter.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/idlejitter.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/idlejitter.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/idlejitter.plugin/metadata.yaml"
sidebar_label: "Idle OS Jitter"
learn_status: "Published"
-learn_rel_path: "Data Collection/Synthetic Checks"
+learn_rel_path: "Collecting Metrics/Synthetic Checks"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -97,7 +97,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/idlejitter.plugin/metadata.yaml b/src/collectors/idlejitter.plugin/metadata.yaml
index 0ad946994..0ad946994 100644
--- a/collectors/idlejitter.plugin/metadata.yaml
+++ b/src/collectors/idlejitter.plugin/metadata.yaml
diff --git a/collectors/idlejitter.plugin/plugin_idlejitter.c b/src/collectors/idlejitter.plugin/plugin_idlejitter.c
index d90548869..f71c99ad1 100644
--- a/collectors/idlejitter.plugin/plugin_idlejitter.c
+++ b/src/collectors/idlejitter.plugin/plugin_idlejitter.c
@@ -68,9 +68,7 @@ void *cpuidlejitter_main(void *ptr) {
usec_t error = dt - sleep_ut;
error_total += error;
- if(unlikely(!iterations))
- error_min = error;
- else if(error < error_min)
+ if(unlikely(!iterations || error < error_min))
error_min = error;
if(error > error_max)
diff --git a/collectors/ioping.plugin/README.md b/src/collectors/ioping.plugin/README.md
index cb660f13b..cb660f13b 120000
--- a/collectors/ioping.plugin/README.md
+++ b/src/collectors/ioping.plugin/README.md
diff --git a/collectors/ioping.plugin/integrations/ioping.md b/src/collectors/ioping.plugin/integrations/ioping.md
index 39a07ed62..b71373e47 100644
--- a/collectors/ioping.plugin/integrations/ioping.md
+++ b/src/collectors/ioping.plugin/integrations/ioping.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/ioping.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/ioping.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/ioping.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/ioping.plugin/metadata.yaml"
sidebar_label: "IOPing"
learn_status: "Published"
-learn_rel_path: "Data Collection/Synthetic Checks"
+learn_rel_path: "Collecting Metrics/Synthetic Checks"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -74,7 +74,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ ioping_disk_latency ](https://github.com/netdata/netdata/blob/master/health/health.d/ioping.conf) | ioping.latency | average I/O latency over the last 10 seconds |
+| [ ioping_disk_latency ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ioping.conf) | ioping.latency | average I/O latency over the last 10 seconds |
## Setup
@@ -95,7 +95,7 @@ The configuration file name for this integration is `ioping.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/ioping.plugin/ioping.conf b/src/collectors/ioping.plugin/ioping.conf
index 86f0de7f4..86f0de7f4 100644
--- a/collectors/ioping.plugin/ioping.conf
+++ b/src/collectors/ioping.plugin/ioping.conf
diff --git a/collectors/ioping.plugin/ioping.plugin.in b/src/collectors/ioping.plugin/ioping.plugin.in
index 171e384db..171e384db 100755
--- a/collectors/ioping.plugin/ioping.plugin.in
+++ b/src/collectors/ioping.plugin/ioping.plugin.in
diff --git a/collectors/log2journal/log2journal-help.c b/src/collectors/log2journal/log2journal-help.c
index 21be948e8..f0536079d 100644
--- a/collectors/log2journal/log2journal-help.c
+++ b/src/collectors/log2journal/log2journal-help.c
@@ -44,7 +44,7 @@ static void config_dir_print_available(void) {
void log_job_command_line_help(const char *name) {
printf("\n");
- printf("Netdata log2journal " PACKAGE_VERSION "\n");
+ printf("Netdata log2journal " VERSION "\n");
printf("\n");
printf("Convert logs to systemd Journal Export Format.\n");
printf("\n");
diff --git a/collectors/log2journal/log2journal-inject.c b/src/collectors/log2journal/log2journal-inject.c
index 45158066b..45158066b 100644
--- a/collectors/log2journal/log2journal-inject.c
+++ b/src/collectors/log2journal/log2journal-inject.c
diff --git a/collectors/log2journal/log2journal-json.c b/src/collectors/log2journal/log2journal-json.c
index 2ca294e4d..be369ffe9 100644
--- a/collectors/log2journal/log2journal-json.c
+++ b/src/collectors/log2journal/log2journal-json.c
@@ -47,7 +47,7 @@ static inline bool json_expect_char_after_white_space(LOG_JSON_STATE *js, const
}
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: character '%c' is not one of the expected characters (%s), at pos %zu",
+ "JSON PARSER: character '%c' is not one of the expected characters (%s), at pos %u",
*s ? *s : '?', expected, js->pos);
return false;
@@ -62,7 +62,7 @@ static inline bool json_parse_null(LOG_JSON_STATE *js) {
}
else {
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: expected 'null', found '%.4s' at position %zu", s, js->pos);
+ "JSON PARSER: expected 'null', found '%.4s' at position %u", s, js->pos);
return false;
}
}
@@ -76,7 +76,7 @@ static inline bool json_parse_true(LOG_JSON_STATE *js) {
}
else {
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: expected 'true', found '%.4s' at position %zu", s, js->pos);
+ "JSON PARSER: expected 'true', found '%.4s' at position %u", s, js->pos);
return false;
}
}
@@ -90,7 +90,7 @@ static inline bool json_parse_false(LOG_JSON_STATE *js) {
}
else {
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: expected 'false', found '%.4s' at position %zu", s, js->pos);
+ "JSON PARSER: expected 'false', found '%.4s' at position %u", s, js->pos);
return false;
}
}
@@ -112,7 +112,7 @@ static inline bool json_parse_number(LOG_JSON_STATE *js) {
// Digits before decimal point
while (*s >= '0' && *s <= '9') {
if (remaining < 2) {
- snprintf(js->msg, sizeof(js->msg), "JSON PARSER: truncated number value at pos %zu", js->pos);
+ snprintf(js->msg, sizeof(js->msg), "JSON PARSER: truncated number value at position %u", js->pos);
return false;
}
*d++ = *s++;
@@ -126,7 +126,7 @@ static inline bool json_parse_number(LOG_JSON_STATE *js) {
while (*s >= '0' && *s <= '9') {
if (remaining < 2) {
- snprintf(js->msg, sizeof(js->msg), "JSON PARSER: truncated fractional part at pos %zu", js->pos);
+ snprintf(js->msg, sizeof(js->msg), "JSON PARSER: truncated fractional part at position %u", js->pos);
return false;
}
*d++ = *s++;
@@ -147,7 +147,7 @@ static inline bool json_parse_number(LOG_JSON_STATE *js) {
while (*s >= '0' && *s <= '9') {
if (remaining < 2) {
- snprintf(js->msg, sizeof(js->msg), "JSON PARSER: truncated exponent at pos %zu", js->pos);
+ snprintf(js->msg, sizeof(js->msg), "JSON PARSER: truncated exponent at position %u", js->pos);
return false;
}
*d++ = *s++;
@@ -162,7 +162,7 @@ static inline bool json_parse_number(LOG_JSON_STATE *js) {
json_process_key_value(js, value, d - value);
return true;
} else {
- snprintf(js->msg, sizeof(js->msg), "JSON PARSER: invalid number format at pos %zu", js->pos);
+ snprintf(js->msg, sizeof(js->msg), "JSON PARSER: invalid number format at position %u", js->pos);
return false;
}
}
@@ -334,7 +334,7 @@ static inline bool json_parse_string(LOG_JSON_STATE *js) {
if(remaining < 2) {
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: truncated string value at pos %zu", js->pos);
+ "JSON PARSER: truncated string value at position %u", js->pos);
return false;
}
else {
@@ -362,7 +362,7 @@ static inline bool json_parse_key_and_push(LOG_JSON_STATE *js) {
if(js->depth >= JSON_DEPTH_MAX - 1) {
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: object too deep, at pos %zu", js->pos);
+ "JSON PARSER: object too deep, at position %u", js->pos);
return false;
}
@@ -392,7 +392,7 @@ static inline bool json_parse_key_and_push(LOG_JSON_STATE *js) {
else {
if(remaining < 2) {
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: key buffer full - keys are too long, at pos %zu", js->pos);
+ "JSON PARSER: key buffer full - keys are too long, at position %u", js->pos);
return false;
}
*d++ = c;
@@ -417,7 +417,7 @@ static inline bool json_parse_key_and_push(LOG_JSON_STATE *js) {
static inline bool json_key_pop(LOG_JSON_STATE *js) {
if(js->depth <= 0) {
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: cannot pop a key at depth %zu, at pos %zu", js->depth, js->pos);
+ "JSON PARSER: cannot pop a key at depth %u, at position %u", js->depth, js->pos);
return false;
}
@@ -465,7 +465,7 @@ static inline bool json_parse_value(LOG_JSON_STATE *js) {
}
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: unexpected character at pos %zu", js->pos);
+ "JSON PARSER: unexpected character at position %u", js->pos);
return false;
}
@@ -491,7 +491,7 @@ static inline bool json_key_index_and_push(LOG_JSON_STATE *js, size_t index) {
while (*t) {
if(remaining < 2) {
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: key buffer full - keys are too long, at pos %zu", js->pos);
+ "JSON PARSER: key buffer full - keys are too long, at position %u", js->pos);
return false;
}
@@ -613,7 +613,7 @@ bool json_parse_document(LOG_JSON_STATE *js, const char *txt) {
if(*s) {
snprintf(js->msg, sizeof(js->msg),
- "JSON PARSER: excess characters found after document is finished, at pos %zu", js->pos);
+ "JSON PARSER: excess characters found after document is finished, at position %u", js->pos);
return false;
}
diff --git a/collectors/log2journal/log2journal-logfmt.c b/src/collectors/log2journal/log2journal-logfmt.c
index 5966cce90..b500f0d82 100644
--- a/collectors/log2journal/log2journal-logfmt.c
+++ b/src/collectors/log2journal/log2journal-logfmt.c
@@ -98,7 +98,7 @@ static inline bool logftm_parse_value(LOGFMT_STATE *lfs) {
if(remaining < 2) {
snprintf(lfs->msg, sizeof(lfs->msg),
- "LOGFMT PARSER: truncated string value at pos %zu", lfs->pos);
+ "LOGFMT PARSER: truncated string value at position %u", lfs->pos);
return false;
}
else {
@@ -114,7 +114,7 @@ static inline bool logftm_parse_value(LOGFMT_STATE *lfs) {
if(quote != '\0') {
if (*s != quote) {
snprintf(lfs->msg, sizeof(lfs->msg),
- "LOGFMT PARSER: missing quote at pos %zu: '%s'",
+ "LOGFMT PARSER: missing quote at position %u: '%s'",
lfs->pos, s);
return false;
}
@@ -150,7 +150,7 @@ static inline bool logfmt_parse_key(LOGFMT_STATE *lfs) {
else {
if(remaining < 2) {
snprintf(lfs->msg, sizeof(lfs->msg),
- "LOGFMT PARSER: key buffer full - keys are too long, at pos %zu", lfs->pos);
+ "LOGFMT PARSER: key buffer full - keys are too long, at position %u", lfs->pos);
return false;
}
*d++ = c;
@@ -165,7 +165,7 @@ static inline bool logfmt_parse_key(LOGFMT_STATE *lfs) {
s = logfmt_current_pos(lfs);
if(*s != '=') {
snprintf(lfs->msg, sizeof(lfs->msg),
- "LOGFMT PARSER: key is missing the equal sign, at pos %zu", lfs->pos);
+ "LOGFMT PARSER: key is missing the equal sign, at position %u", lfs->pos);
return false;
}
diff --git a/collectors/log2journal/log2journal-params.c b/src/collectors/log2journal/log2journal-params.c
index a7bb3e263..a7bb3e263 100644
--- a/collectors/log2journal/log2journal-params.c
+++ b/src/collectors/log2journal/log2journal-params.c
diff --git a/collectors/log2journal/log2journal-pattern.c b/src/collectors/log2journal/log2journal-pattern.c
index 4b7e9026b..4b7e9026b 100644
--- a/collectors/log2journal/log2journal-pattern.c
+++ b/src/collectors/log2journal/log2journal-pattern.c
diff --git a/collectors/log2journal/log2journal-pcre2.c b/src/collectors/log2journal/log2journal-pcre2.c
index 185e69108..185e69108 100644
--- a/collectors/log2journal/log2journal-pcre2.c
+++ b/src/collectors/log2journal/log2journal-pcre2.c
diff --git a/collectors/log2journal/log2journal-rename.c b/src/collectors/log2journal/log2journal-rename.c
index c6975779f..c6975779f 100644
--- a/collectors/log2journal/log2journal-rename.c
+++ b/src/collectors/log2journal/log2journal-rename.c
diff --git a/collectors/log2journal/log2journal-replace.c b/src/collectors/log2journal/log2journal-replace.c
index 429d615da..7075d109d 100644
--- a/collectors/log2journal/log2journal-replace.c
+++ b/src/collectors/log2journal/log2journal-replace.c
@@ -74,6 +74,7 @@ bool replace_pattern_set(REPLACE_PATTERN *rp, const char *pattern) {
log2stderr("Error: Failed to add replacement node for variable.");
return false;
}
+ freez(variable_name);
current = end + 1; // Move past the variable
}
@@ -97,6 +98,7 @@ bool replace_pattern_set(REPLACE_PATTERN *rp, const char *pattern) {
log2stderr("Error: Failed to add replacement node for text.");
return false;
}
+ freez(text);
}
}
diff --git a/collectors/log2journal/log2journal-rewrite.c b/src/collectors/log2journal/log2journal-rewrite.c
index 112391bf0..112391bf0 100644
--- a/collectors/log2journal/log2journal-rewrite.c
+++ b/src/collectors/log2journal/log2journal-rewrite.c
diff --git a/collectors/log2journal/log2journal-yaml.c b/src/collectors/log2journal/log2journal-yaml.c
index 862e7bf4b..cfdbcf4bd 100644
--- a/collectors/log2journal/log2journal-yaml.c
+++ b/src/collectors/log2journal/log2journal-yaml.c
@@ -464,6 +464,7 @@ static size_t yaml_parse_rewrites(yaml_parser_t *parser, LOG_JOB *jb) {
yaml_error(parser, &sub_event, "Expected scalar for rewrite key");
errors++;
} else {
+ freez(key);
key = strndupz((char *)sub_event.data.scalar.value, sub_event.data.scalar.length);
yaml_event_delete(&sub_event);
}
@@ -498,6 +499,7 @@ static size_t yaml_parse_rewrites(yaml_parser_t *parser, LOG_JOB *jb) {
yaml_error(parser, &sub_event, "Expected scalar for rewrite value");
errors++;
} else {
+ freez(replace_pattern);
replace_pattern = strndupz((char *)sub_event.data.scalar.value, sub_event.data.scalar.length);
yaml_event_delete(&sub_event);
}
@@ -561,6 +563,12 @@ static size_t yaml_parse_rewrites(yaml_parser_t *parser, LOG_JOB *jb) {
yaml_event_delete(&sub_event);
}
+ freez(replace_pattern);
+ replace_pattern = NULL;
+ freez(search_pattern);
+ search_pattern = NULL;
+ freez(key);
+ key = NULL;
}
break;
diff --git a/collectors/log2journal/log2journal.c b/src/collectors/log2journal/log2journal.c
index c3204939c..e4de6cd24 100644
--- a/collectors/log2journal/log2journal.c
+++ b/src/collectors/log2journal/log2journal.c
@@ -67,7 +67,7 @@ static inline HASHED_KEY *get_key_from_hashtable(LOG_JOB *jb, HASHED_KEY *k) {
if(!k->hashtable_ptr) {
HASHED_KEY *ht_key;
- SIMPLE_HASHTABLE_SLOT_KEY *slot = simple_hashtable_get_slot_KEY(&jb->hashtable, k->hash, true);
+ SIMPLE_HASHTABLE_SLOT_KEY *slot = simple_hashtable_get_slot_KEY(&jb->hashtable, k->hash, NULL, true);
if((ht_key = SIMPLE_HASHTABLE_SLOT_DATA(slot))) {
if(!(ht_key->flags & HK_COLLISION_CHECKED)) {
ht_key->flags |= HK_COLLISION_CHECKED;
diff --git a/collectors/log2journal/log2journal.d/default.yaml b/src/collectors/log2journal/log2journal.d/default.yaml
index d41efc4ab..d41efc4ab 100644
--- a/collectors/log2journal/log2journal.d/default.yaml
+++ b/src/collectors/log2journal/log2journal.d/default.yaml
diff --git a/collectors/log2journal/log2journal.d/nginx-combined.yaml b/src/collectors/log2journal/log2journal.d/nginx-combined.yaml
index 003c774d7..003c774d7 100644
--- a/collectors/log2journal/log2journal.d/nginx-combined.yaml
+++ b/src/collectors/log2journal/log2journal.d/nginx-combined.yaml
diff --git a/collectors/log2journal/log2journal.d/nginx-json.yaml b/src/collectors/log2journal/log2journal.d/nginx-json.yaml
index 7fdc4be58..7fdc4be58 100644
--- a/collectors/log2journal/log2journal.d/nginx-json.yaml
+++ b/src/collectors/log2journal/log2journal.d/nginx-json.yaml
diff --git a/collectors/log2journal/log2journal.h b/src/collectors/log2journal/log2journal.h
index 834a5b135..c52e0898e 100644
--- a/collectors/log2journal/log2journal.h
+++ b/src/collectors/log2journal/log2journal.h
@@ -4,7 +4,7 @@
#define NETDATA_LOG2JOURNAL_H
// only for PACKAGE_VERSION
-#include "config.h"
+#include <config.h>
#include <stdio.h>
#include <stdlib.h>
@@ -86,7 +86,7 @@ static inline void freez(void *ptr) {
// ----------------------------------------------------------------------------
#define XXH_INLINE_ALL
-#include "../../libnetdata/xxhash.h"
+#include "libnetdata/xxhash.h"
#define PCRE2_CODE_UNIT_WIDTH 8
#include <pcre2.h>
@@ -99,17 +99,14 @@ static inline void freez(void *ptr) {
// hashtable for HASHED_KEY
// cleanup hashtable defines
-#undef SIMPLE_HASHTABLE_SORT_FUNCTION
-#undef SIMPLE_HASHTABLE_VALUE_TYPE
-#undef SIMPLE_HASHTABLE_NAME
-#undef NETDATA_SIMPLE_HASHTABLE_H
+#include "libnetdata/simple_hashtable_undef.h"
struct hashed_key;
static inline int compare_keys(struct hashed_key *k1, struct hashed_key *k2);
#define SIMPLE_HASHTABLE_SORT_FUNCTION compare_keys
#define SIMPLE_HASHTABLE_VALUE_TYPE struct hashed_key
#define SIMPLE_HASHTABLE_NAME _KEY
-#include "../../libnetdata/simple_hashtable.h"
+#include "libnetdata/simple_hashtable.h"
// ----------------------------------------------------------------------------
diff --git a/collectors/log2journal/tests.d/default.output b/src/collectors/log2journal/tests.d/default.output
index ef17cb2c7..ef17cb2c7 100644
--- a/collectors/log2journal/tests.d/default.output
+++ b/src/collectors/log2journal/tests.d/default.output
diff --git a/collectors/log2journal/tests.d/full.output b/src/collectors/log2journal/tests.d/full.output
index 074092d4e..074092d4e 100644
--- a/collectors/log2journal/tests.d/full.output
+++ b/src/collectors/log2journal/tests.d/full.output
diff --git a/collectors/log2journal/tests.d/full.yaml b/src/collectors/log2journal/tests.d/full.yaml
index 86cafb5a2..86cafb5a2 100644
--- a/collectors/log2journal/tests.d/full.yaml
+++ b/src/collectors/log2journal/tests.d/full.yaml
diff --git a/collectors/log2journal/tests.d/json-exclude.output b/src/collectors/log2journal/tests.d/json-exclude.output
index a8f6f83e6..a8f6f83e6 100644
--- a/collectors/log2journal/tests.d/json-exclude.output
+++ b/src/collectors/log2journal/tests.d/json-exclude.output
diff --git a/collectors/log2journal/tests.d/json-include.output b/src/collectors/log2journal/tests.d/json-include.output
index 326c58da2..326c58da2 100644
--- a/collectors/log2journal/tests.d/json-include.output
+++ b/src/collectors/log2journal/tests.d/json-include.output
diff --git a/collectors/log2journal/tests.d/json.log b/src/collectors/log2journal/tests.d/json.log
index 3f1334960..3f1334960 100644
--- a/collectors/log2journal/tests.d/json.log
+++ b/src/collectors/log2journal/tests.d/json.log
diff --git a/collectors/log2journal/tests.d/json.output b/src/collectors/log2journal/tests.d/json.output
index 83499cc55..83499cc55 100644
--- a/collectors/log2journal/tests.d/json.output
+++ b/src/collectors/log2journal/tests.d/json.output
diff --git a/collectors/log2journal/tests.d/logfmt.log b/src/collectors/log2journal/tests.d/logfmt.log
index e55a83bbb..e55a83bbb 100644
--- a/collectors/log2journal/tests.d/logfmt.log
+++ b/src/collectors/log2journal/tests.d/logfmt.log
diff --git a/collectors/log2journal/tests.d/logfmt.output b/src/collectors/log2journal/tests.d/logfmt.output
index 4291c9665..4291c9665 100644
--- a/collectors/log2journal/tests.d/logfmt.output
+++ b/src/collectors/log2journal/tests.d/logfmt.output
diff --git a/collectors/log2journal/tests.d/logfmt.yaml b/src/collectors/log2journal/tests.d/logfmt.yaml
index 91e93a71e..91e93a71e 100644
--- a/collectors/log2journal/tests.d/logfmt.yaml
+++ b/src/collectors/log2journal/tests.d/logfmt.yaml
diff --git a/collectors/log2journal/tests.d/nginx-combined.log b/src/collectors/log2journal/tests.d/nginx-combined.log
index b0faa81e9..b0faa81e9 100644
--- a/collectors/log2journal/tests.d/nginx-combined.log
+++ b/src/collectors/log2journal/tests.d/nginx-combined.log
diff --git a/collectors/log2journal/tests.d/nginx-combined.output b/src/collectors/log2journal/tests.d/nginx-combined.output
index 07fd11014..07fd11014 100644
--- a/collectors/log2journal/tests.d/nginx-combined.output
+++ b/src/collectors/log2journal/tests.d/nginx-combined.output
diff --git a/collectors/log2journal/tests.d/nginx-json.log b/src/collectors/log2journal/tests.d/nginx-json.log
index 7e2b5d5f5..7e2b5d5f5 100644
--- a/collectors/log2journal/tests.d/nginx-json.log
+++ b/src/collectors/log2journal/tests.d/nginx-json.log
diff --git a/collectors/log2journal/tests.d/nginx-json.output b/src/collectors/log2journal/tests.d/nginx-json.output
index e7db9dcbd..e7db9dcbd 100644
--- a/collectors/log2journal/tests.d/nginx-json.output
+++ b/src/collectors/log2journal/tests.d/nginx-json.output
diff --git a/collectors/log2journal/tests.sh b/src/collectors/log2journal/tests.sh
index 402438866..402438866 100755
--- a/collectors/log2journal/tests.sh
+++ b/src/collectors/log2journal/tests.sh
diff --git a/collectors/macos.plugin/README.md b/src/collectors/macos.plugin/README.md
index 2ea6842e4..2ea6842e4 120000
--- a/collectors/macos.plugin/README.md
+++ b/src/collectors/macos.plugin/README.md
diff --git a/collectors/macos.plugin/integrations/macos.md b/src/collectors/macos.plugin/integrations/macos.md
index 5128a5a77..d87c7910d 100644
--- a/collectors/macos.plugin/integrations/macos.md
+++ b/src/collectors/macos.plugin/integrations/macos.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/macos.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/macos.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/macos.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/macos.plugin/metadata.yaml"
sidebar_label: "macOS"
learn_status: "Published"
-learn_rel_path: "Data Collection/macOS Systems"
+learn_rel_path: "Collecting Metrics/macOS Systems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -163,7 +163,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ interface_speed ](https://github.com/netdata/netdata/blob/master/health/health.d/net.conf) | net.net | network interface ${label:device} current speed |
+| [ interface_speed ](https://github.com/netdata/netdata/blob/master/src/health/health.d/net.conf) | net.net | network interface ${label:device} current speed |
## Setup
@@ -189,7 +189,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/macos.plugin/macos_fw.c b/src/collectors/macos.plugin/macos_fw.c
index 75ef386b9..75ef386b9 100644
--- a/collectors/macos.plugin/macos_fw.c
+++ b/src/collectors/macos.plugin/macos_fw.c
diff --git a/collectors/macos.plugin/macos_mach_smi.c b/src/collectors/macos.plugin/macos_mach_smi.c
index 30c957187..30c957187 100644
--- a/collectors/macos.plugin/macos_mach_smi.c
+++ b/src/collectors/macos.plugin/macos_mach_smi.c
diff --git a/collectors/macos.plugin/macos_sysctl.c b/src/collectors/macos.plugin/macos_sysctl.c
index 520d2f938..520d2f938 100644
--- a/collectors/macos.plugin/macos_sysctl.c
+++ b/src/collectors/macos.plugin/macos_sysctl.c
diff --git a/collectors/macos.plugin/plugin_macos.c b/src/collectors/macos.plugin/plugin_macos.c
index 3aaa46c72..3aaa46c72 100644
--- a/collectors/macos.plugin/plugin_macos.c
+++ b/src/collectors/macos.plugin/plugin_macos.c
diff --git a/collectors/macos.plugin/plugin_macos.h b/src/collectors/macos.plugin/plugin_macos.h
index 2c673a224..2c673a224 100644
--- a/collectors/macos.plugin/plugin_macos.h
+++ b/src/collectors/macos.plugin/plugin_macos.h
diff --git a/collectors/nfacct.plugin/README.md b/src/collectors/nfacct.plugin/README.md
index ea320d139..ea320d139 120000
--- a/collectors/nfacct.plugin/README.md
+++ b/src/collectors/nfacct.plugin/README.md
diff --git a/collectors/nfacct.plugin/integrations/netfilter.md b/src/collectors/nfacct.plugin/integrations/netfilter.md
index 831b6fb5b..c36b5e999 100644
--- a/collectors/nfacct.plugin/integrations/netfilter.md
+++ b/src/collectors/nfacct.plugin/integrations/netfilter.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/nfacct.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/nfacct.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/nfacct.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/nfacct.plugin/metadata.yaml"
sidebar_label: "Netfilter"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Firewall"
+learn_rel_path: "Collecting Metrics/Linux Systems/Firewall"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -107,7 +107,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/nfacct.plugin/metadata.yaml b/src/collectors/nfacct.plugin/metadata.yaml
index 943471a37..943471a37 100644
--- a/collectors/nfacct.plugin/metadata.yaml
+++ b/src/collectors/nfacct.plugin/metadata.yaml
diff --git a/collectors/nfacct.plugin/plugin_nfacct.c b/src/collectors/nfacct.plugin/plugin_nfacct.c
index 2863cd7eb..1d46f2953 100644
--- a/collectors/nfacct.plugin/plugin_nfacct.c
+++ b/src/collectors/nfacct.plugin/plugin_nfacct.c
@@ -796,7 +796,7 @@ int main(int argc, char **argv) {
" --help print this message and exit\n"
"\n"
" For more information:\n"
- " https://github.com/netdata/netdata/tree/master/collectors/nfacct.plugin\n"
+ " https://github.com/netdata/netdata/tree/master/src/collectors/nfacct.plugin\n"
"\n"
, VERSION
, netdata_update_every
diff --git a/collectors/perf.plugin/README.md b/src/collectors/perf.plugin/README.md
index fb8a0cd69..fb8a0cd69 120000
--- a/collectors/perf.plugin/README.md
+++ b/src/collectors/perf.plugin/README.md
diff --git a/collectors/perf.plugin/integrations/cpu_performance.md b/src/collectors/perf.plugin/integrations/cpu_performance.md
index d3c316d2e..93df00dd8 100644
--- a/collectors/perf.plugin/integrations/cpu_performance.md
+++ b/src/collectors/perf.plugin/integrations/cpu_performance.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/perf.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/perf.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/perf.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/perf.plugin/metadata.yaml"
sidebar_label: "CPU performance"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems"
+learn_rel_path: "Collecting Metrics/Linux Systems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -104,14 +104,14 @@ If you are [using our official native DEB/RPM packages](https://github.com/netda
The plugin is disabled by default because the number of PMUs is usually quite limited and it is not desired to allow Netdata to struggle silently for PMUs, interfering with other performance monitoring software.
-To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`, to edit the `netdata.conf` file.
+To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md), which is typically at `/etc/netdata`, to edit the `netdata.conf` file.
```bash
cd /etc/netdata # Replace this path with your Netdata config directory, if different
sudo ./edit-config netdata.conf
```
-Change the value of the `perf` setting to `yes` in the `[plugins]` section. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
+Change the value of the `perf` setting to `yes` in the `[plugins]` section. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/packaging/installer/README.md#maintaining-a-netdata-agent-installation) for your system.
@@ -133,7 +133,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/perf.plugin/perf_plugin.c b/src/collectors/perf.plugin/perf_plugin.c
index fe3b04daa..b839263d1 100644
--- a/collectors/perf.plugin/perf_plugin.c
+++ b/src/collectors/perf.plugin/perf_plugin.c
@@ -35,8 +35,8 @@
#define RRD_FAMILY_SW "software"
#define RRD_FAMILY_CACHE "cache"
-#define NO_FD -1
-#define ALL_PIDS -1
+#define NO_FD (-1)
+#define ALL_PIDS (-1)
#define RUNNING_THRESHOLD 100
static int debug = 0;
@@ -1264,7 +1264,7 @@ void parse_command_line(int argc, char **argv) {
" --help print this message and exit\n"
"\n"
" For more information:\n"
- " https://github.com/netdata/netdata/tree/master/collectors/perf.plugin\n"
+ " https://github.com/netdata/netdata/tree/master/src/collectors/perf.plugin\n"
"\n"
, VERSION
, update_every
diff --git a/collectors/plugins.d/plugins_d.c b/src/collectors/plugins.d/plugins_d.c
index 20061ad29..0bcb3df63 100644
--- a/collectors/plugins.d/plugins_d.c
+++ b/src/collectors/plugins.d/plugins_d.c
@@ -140,67 +140,65 @@ static void pluginsd_worker_thread_handle_error(struct plugind *cd, int worker_r
static void *pluginsd_worker_thread(void *arg) {
worker_register("PLUGINSD");
- netdata_thread_cleanup_push(pluginsd_worker_thread_cleanup, arg);
+ netdata_thread_cleanup_push(pluginsd_worker_thread_cleanup, arg)
+ {
+ struct plugind *cd = (struct plugind *) arg;
+ plugin_set_running(cd);
- {
- struct plugind *cd = (struct plugind *) arg;
- plugin_set_running(cd);
+ size_t count = 0;
- size_t count = 0;
+ while(service_running(SERVICE_COLLECTORS)) {
+ FILE *fp_child_input = NULL;
+ FILE *fp_child_output = netdata_popen(cd->cmd, &cd->unsafe.pid, &fp_child_input);
- while(service_running(SERVICE_COLLECTORS)) {
- FILE *fp_child_input = NULL;
- FILE *fp_child_output = netdata_popen(cd->cmd, &cd->unsafe.pid, &fp_child_input);
-
- if(unlikely(!fp_child_input || !fp_child_output)) {
- netdata_log_error("PLUGINSD: 'host:%s', cannot popen(\"%s\", \"r\").",
- rrdhost_hostname(cd->host), cd->cmd);
- break;
- }
-
- nd_log(NDLS_DAEMON, NDLP_DEBUG,
- "PLUGINSD: 'host:%s' connected to '%s' running on pid %d",
- rrdhost_hostname(cd->host),
- cd->fullfilename, cd->unsafe.pid);
+ if(unlikely(!fp_child_input || !fp_child_output)) {
+ netdata_log_error("PLUGINSD: 'host:%s', cannot popen(\"%s\", \"r\").",
+ rrdhost_hostname(cd->host), cd->cmd);
+ break;
+ }
- const char *plugin = strrchr(cd->fullfilename, '/');
- if(plugin)
- plugin++;
- else
- plugin = cd->fullfilename;
+ nd_log(NDLS_DAEMON, NDLP_DEBUG,
+ "PLUGINSD: 'host:%s' connected to '%s' running on pid %d",
+ rrdhost_hostname(cd->host),
+ cd->fullfilename, cd->unsafe.pid);
- char module[100];
- snprintfz(module, sizeof(module), "plugins.d[%s]", plugin);
- ND_LOG_STACK lgs[] = {
- ND_LOG_FIELD_TXT(NDF_MODULE, module),
- ND_LOG_FIELD_TXT(NDF_NIDL_NODE, rrdhost_hostname(cd->host)),
- ND_LOG_FIELD_TXT(NDF_SRC_TRANSPORT, "pluginsd"),
- ND_LOG_FIELD_END(),
- };
- ND_LOG_STACK_PUSH(lgs);
+ const char *plugin = strrchr(cd->fullfilename, '/');
+ if(plugin)
+ plugin++;
+ else
+ plugin = cd->fullfilename;
- count = pluginsd_process(cd->host, cd, fp_child_input, fp_child_output, 0);
+ char module[100];
+ snprintfz(module, sizeof(module), "plugins.d[%s]", plugin);
+ ND_LOG_STACK lgs[] = {
+ ND_LOG_FIELD_TXT(NDF_MODULE, module),
+ ND_LOG_FIELD_TXT(NDF_NIDL_NODE, rrdhost_hostname(cd->host)),
+ ND_LOG_FIELD_TXT(NDF_SRC_TRANSPORT, "pluginsd"),
+ ND_LOG_FIELD_END(),
+ };
+ ND_LOG_STACK_PUSH(lgs);
- nd_log(NDLS_DAEMON, NDLP_DEBUG,
- "PLUGINSD: 'host:%s', '%s' (pid %d) disconnected after %zu successful data collections (ENDs).",
- rrdhost_hostname(cd->host), cd->fullfilename, cd->unsafe.pid, count);
+ count = pluginsd_process(cd->host, cd, fp_child_input, fp_child_output, 0);
- killpid(cd->unsafe.pid);
+ nd_log(NDLS_DAEMON, NDLP_DEBUG,
+ "PLUGINSD: 'host:%s', '%s' (pid %d) disconnected after %zu successful data collections (ENDs).",
+ rrdhost_hostname(cd->host), cd->fullfilename, cd->unsafe.pid, count);
- int worker_ret_code = netdata_pclose(fp_child_input, fp_child_output, cd->unsafe.pid);
+ killpid(cd->unsafe.pid);
- if(likely(worker_ret_code == 0))
- pluginsd_worker_thread_handle_success(cd);
- else
- pluginsd_worker_thread_handle_error(cd, worker_ret_code);
+ int worker_ret_code = netdata_pclose(fp_child_input, fp_child_output, cd->unsafe.pid);
- cd->unsafe.pid = 0;
+ if(likely(worker_ret_code == 0))
+ pluginsd_worker_thread_handle_success(cd);
+ else
+ pluginsd_worker_thread_handle_error(cd, worker_ret_code);
- if(unlikely(!plugin_is_enabled(cd)))
- break;
- }
- }
+ cd->unsafe.pid = 0;
+ if(unlikely(!plugin_is_enabled(cd)))
+ break;
+ }
+ }
netdata_thread_cleanup_pop(1);
return NULL;
}
diff --git a/collectors/plugins.d/pluginsd_parser.h b/src/collectors/plugins.d/pluginsd_parser.h
index 1fce9a89a..d317a77be 100644
--- a/collectors/plugins.d/pluginsd_parser.h
+++ b/src/collectors/plugins.d/pluginsd_parser.h
@@ -112,7 +112,7 @@ typedef struct parser {
struct buffered_reader reader;
struct line_splitter line;
- PARSER_KEYWORD *keyword;
+ const PARSER_KEYWORD *keyword;
struct {
const char *end_keyword;
@@ -123,7 +123,7 @@ typedef struct parser {
struct {
DICTIONARY *functions;
- usec_t smaller_timeout;
+ usec_t smaller_monotonic_timeout_ut;
} inflight;
struct {
@@ -136,9 +136,8 @@ PARSER *parser_init(struct parser_user_object *user, FILE *fp_input, FILE *fp_ou
void parser_init_repertoire(PARSER *parser, PARSER_REPERTOIRE repertoire);
void parser_destroy(PARSER *working_parser);
void pluginsd_cleanup_v2(PARSER *parser);
-void inflight_functions_init(PARSER *parser);
void pluginsd_keywords_init(PARSER *parser, PARSER_REPERTOIRE repertoire);
-PARSER_RC parser_execute(PARSER *parser, PARSER_KEYWORD *keyword, char **words, size_t num_words);
+PARSER_RC parser_execute(PARSER *parser, const PARSER_KEYWORD *keyword, char **words, size_t num_words);
static inline int find_first_keyword(const char *src, char *dst, int dst_size, bool *isspace_map) {
const char *s = src, *keyword_start;
@@ -154,10 +153,10 @@ static inline int find_first_keyword(const char *src, char *dst, int dst_size, b
return dst_size == 0 ? 0 : (int) (s - keyword_start);
}
-PARSER_KEYWORD *gperf_lookup_keyword(register const char *str, register size_t len);
+const PARSER_KEYWORD *gperf_lookup_keyword(register const char *str, register size_t len);
-static inline PARSER_KEYWORD *parser_find_keyword(PARSER *parser, const char *command) {
- PARSER_KEYWORD *t = gperf_lookup_keyword(command, strlen(command));
+static inline const PARSER_KEYWORD *parser_find_keyword(PARSER *parser, const char *command) {
+ const PARSER_KEYWORD *t = gperf_lookup_keyword(command, strlen(command));
if(t && (t->repertoire & parser->repertoire))
return t;
@@ -186,7 +185,7 @@ static inline int parser_action(PARSER *parser, char *input) {
if(buffer_strlen(parser->defer.response) > PLUGINSD_MAX_DEFERRED_SIZE) {
// more than PLUGINSD_MAX_DEFERRED_SIZE of data,
// or a bad plugin that did not send the end_keyword
- internal_error(true, "PLUGINSD: deferred response is too big (%zu bytes). Stopping this plugin.", buffer_strlen(parser->defer.response));
+ nd_log(NDLS_DAEMON, NDLP_ERR, "PLUGINSD: deferred response is too big (%zu bytes). Stopping this plugin.", buffer_strlen(parser->defer.response));
return 1;
}
}
@@ -232,7 +231,7 @@ static inline int parser_action(PARSER *parser, char *input) {
rc = PARSER_RC_ERROR;
if(rc == PARSER_RC_ERROR) {
- CLEAN_BUFFER *wb = buffer_create(PLUGINSD_LINE_MAX, NULL);
+ CLEAN_BUFFER *wb = buffer_create(1024, NULL);
line_splitter_reconstruct_line(wb, &parser->line);
netdata_log_error("PLUGINSD: parser_action('%s') failed on line %zu: { %s } (quotes added to show parsing)",
command, parser->line.count, buffer_tostring(wb));
diff --git a/collectors/proc.plugin/integrations/amd_gpu.md b/src/collectors/proc.plugin/integrations/amd_gpu.md
index e85cce221..24f480894 100644
--- a/collectors/proc.plugin/integrations/amd_gpu.md
+++ b/src/collectors/proc.plugin/integrations/amd_gpu.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/amd_gpu.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/amd_gpu.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "AMD GPU"
learn_status: "Published"
-learn_rel_path: "Data Collection/Hardware Devices and Sensors"
+learn_rel_path: "Collecting Metrics/Hardware Devices and Sensors"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/conntrack.md b/src/collectors/proc.plugin/integrations/conntrack.md
index b38f6b508..33f11db24 100644
--- a/collectors/proc.plugin/integrations/conntrack.md
+++ b/src/collectors/proc.plugin/integrations/conntrack.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/conntrack.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/conntrack.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Conntrack"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Firewall"
+learn_rel_path: "Collecting Metrics/Linux Systems/Firewall"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -79,7 +79,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ netfilter_conntrack_full ](https://github.com/netdata/netdata/blob/master/health/health.d/netfilter.conf) | netfilter.conntrack_sockets | netfilter connection tracker table size utilization |
+| [ netfilter_conntrack_full ](https://github.com/netdata/netdata/blob/master/src/health/health.d/netfilter.conf) | netfilter.conntrack_sockets | netfilter connection tracker table size utilization |
## Setup
diff --git a/collectors/proc.plugin/integrations/disk_statistics.md b/src/collectors/proc.plugin/integrations/disk_statistics.md
index 8f7448c39..9dcfa2ede 100644
--- a/collectors/proc.plugin/integrations/disk_statistics.md
+++ b/src/collectors/proc.plugin/integrations/disk_statistics.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/disk_statistics.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/disk_statistics.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Disk Statistics"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Disk"
+learn_rel_path: "Collecting Metrics/Linux Systems/Disk"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -120,10 +120,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 10min_disk_backlog ](https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf) | disk.backlog | average backlog size of the ${label:device} disk over the last 10 minutes |
-| [ 10min_disk_utilization ](https://github.com/netdata/netdata/blob/master/health/health.d/disks.conf) | disk.util | average percentage of time ${label:device} disk was busy over the last 10 minutes |
-| [ bcache_cache_dirty ](https://github.com/netdata/netdata/blob/master/health/health.d/bcache.conf) | disk.bcache_cache_alloc | percentage of cache space used for dirty data and metadata (this usually means your SSD cache is too small) |
-| [ bcache_cache_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/bcache.conf) | disk.bcache_cache_read_races | number of times data was read from the cache, the bucket was reused and invalidated in the last 10 minutes (when this occurs the data is reread from the backing device) |
+| [ 10min_disk_backlog ](https://github.com/netdata/netdata/blob/master/src/health/health.d/disks.conf) | disk.backlog | average backlog size of the ${label:device} disk over the last 10 minutes |
+| [ 10min_disk_utilization ](https://github.com/netdata/netdata/blob/master/src/health/health.d/disks.conf) | disk.util | average percentage of time ${label:device} disk was busy over the last 10 minutes |
+| [ bcache_cache_dirty ](https://github.com/netdata/netdata/blob/master/src/health/health.d/bcache.conf) | disk.bcache_cache_alloc | percentage of cache space used for dirty data and metadata (this usually means your SSD cache is too small) |
+| [ bcache_cache_errors ](https://github.com/netdata/netdata/blob/master/src/health/health.d/bcache.conf) | disk.bcache_cache_read_races | number of times data was read from the cache, the bucket was reused and invalidated in the last 10 minutes (when this occurs the data is reread from the backing device) |
## Setup
diff --git a/collectors/proc.plugin/integrations/entropy.md b/src/collectors/proc.plugin/integrations/entropy.md
index 8432a1f96..03b51ecc8 100644
--- a/collectors/proc.plugin/integrations/entropy.md
+++ b/src/collectors/proc.plugin/integrations/entropy.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/entropy.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/entropy.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Entropy"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/System"
+learn_rel_path: "Collecting Metrics/Linux Systems/System"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -107,7 +107,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ lowest_entropy ](https://github.com/netdata/netdata/blob/master/health/health.d/entropy.conf) | system.entropy | minimum number of bits of entropy available for the kernel’s random number generator |
+| [ lowest_entropy ](https://github.com/netdata/netdata/blob/master/src/health/health.d/entropy.conf) | system.entropy | minimum number of bits of entropy available for the kernel’s random number generator |
## Setup
diff --git a/collectors/proc.plugin/integrations/infiniband.md b/src/collectors/proc.plugin/integrations/infiniband.md
index 6cb5fdc53..5a4f5d702 100644
--- a/collectors/proc.plugin/integrations/infiniband.md
+++ b/src/collectors/proc.plugin/integrations/infiniband.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/infiniband.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/infiniband.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "InfiniBand"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
+learn_rel_path: "Collecting Metrics/Linux Systems/Network"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/inter_process_communication.md b/src/collectors/proc.plugin/integrations/inter_process_communication.md
index 55708a4b0..363fbea41 100644
--- a/collectors/proc.plugin/integrations/inter_process_communication.md
+++ b/src/collectors/proc.plugin/integrations/inter_process_communication.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/inter_process_communication.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/inter_process_communication.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Inter Process Communication"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/IPC"
+learn_rel_path: "Collecting Metrics/Linux Systems/IPC"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -93,8 +93,8 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ semaphores_used ](https://github.com/netdata/netdata/blob/master/health/health.d/ipc.conf) | system.ipc_semaphores | IPC semaphore utilization |
-| [ semaphore_arrays_used ](https://github.com/netdata/netdata/blob/master/health/health.d/ipc.conf) | system.ipc_semaphore_arrays | IPC semaphore arrays utilization |
+| [ semaphores_used ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ipc.conf) | system.ipc_semaphores | IPC semaphore utilization |
+| [ semaphore_arrays_used ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ipc.conf) | system.ipc_semaphore_arrays | IPC semaphore arrays utilization |
## Setup
diff --git a/collectors/proc.plugin/integrations/interrupts.md b/src/collectors/proc.plugin/integrations/interrupts.md
index 1b85fb767..b0d39dbbe 100644
--- a/collectors/proc.plugin/integrations/interrupts.md
+++ b/src/collectors/proc.plugin/integrations/interrupts.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/interrupts.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/interrupts.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Interrupts"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/CPU"
+learn_rel_path: "Collecting Metrics/Linux Systems/CPU"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/ip_virtual_server.md b/src/collectors/proc.plugin/integrations/ip_virtual_server.md
index 5c7afd2eb..974c2f60c 100644
--- a/collectors/proc.plugin/integrations/ip_virtual_server.md
+++ b/src/collectors/proc.plugin/integrations/ip_virtual_server.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/ip_virtual_server.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/ip_virtual_server.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "IP Virtual Server"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
+learn_rel_path: "Collecting Metrics/Linux Systems/Network"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/ipv6_socket_statistics.md b/src/collectors/proc.plugin/integrations/ipv6_socket_statistics.md
index 2c1ee2721..0840d3f0f 100644
--- a/collectors/proc.plugin/integrations/ipv6_socket_statistics.md
+++ b/src/collectors/proc.plugin/integrations/ipv6_socket_statistics.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/ipv6_socket_statistics.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/ipv6_socket_statistics.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "IPv6 Socket Statistics"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
+learn_rel_path: "Collecting Metrics/Linux Systems/Network"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/kernel_same-page_merging.md b/src/collectors/proc.plugin/integrations/kernel_same-page_merging.md
index 336f0feaf..37b64d253 100644
--- a/collectors/proc.plugin/integrations/kernel_same-page_merging.md
+++ b/src/collectors/proc.plugin/integrations/kernel_same-page_merging.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/kernel_same-page_merging.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/kernel_same-page_merging.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Kernel Same-Page Merging"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Memory"
+learn_rel_path: "Collecting Metrics/Linux Systems/Memory"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/md_raid.md b/src/collectors/proc.plugin/integrations/md_raid.md
index 34a4840bb..f96f4c5b1 100644
--- a/collectors/proc.plugin/integrations/md_raid.md
+++ b/src/collectors/proc.plugin/integrations/md_raid.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/md_raid.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/md_raid.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "MD RAID"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Disk"
+learn_rel_path: "Collecting Metrics/Linux Systems/Disk"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -96,10 +96,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ mdstat_last_collected ](https://github.com/netdata/netdata/blob/master/health/health.d/mdstat.conf) | md.disks | number of seconds since the last successful data collection |
-| [ mdstat_disks ](https://github.com/netdata/netdata/blob/master/health/health.d/mdstat.conf) | md.disks | number of devices in the down state for the ${label:device} ${label:raid_level} array. Any number > 0 indicates that the array is degraded. |
-| [ mdstat_mismatch_cnt ](https://github.com/netdata/netdata/blob/master/health/health.d/mdstat.conf) | md.mismatch_cnt | number of unsynchronized blocks for the ${label:device} ${label:raid_level} array |
-| [ mdstat_nonredundant_last_collected ](https://github.com/netdata/netdata/blob/master/health/health.d/mdstat.conf) | md.nonredundant | number of seconds since the last successful data collection |
+| [ mdstat_last_collected ](https://github.com/netdata/netdata/blob/master/src/health/health.d/mdstat.conf) | md.disks | number of seconds since the last successful data collection |
+| [ mdstat_disks ](https://github.com/netdata/netdata/blob/master/src/health/health.d/mdstat.conf) | md.disks | number of devices in the down state for the ${label:device} ${label:raid_level} array. Any number > 0 indicates that the array is degraded. |
+| [ mdstat_mismatch_cnt ](https://github.com/netdata/netdata/blob/master/src/health/health.d/mdstat.conf) | md.mismatch_cnt | number of unsynchronized blocks for the ${label:device} ${label:raid_level} array |
+| [ mdstat_nonredundant_last_collected ](https://github.com/netdata/netdata/blob/master/src/health/health.d/mdstat.conf) | md.nonredundant | number of seconds since the last successful data collection |
## Setup
diff --git a/collectors/proc.plugin/integrations/memory_modules_dimms.md b/src/collectors/proc.plugin/integrations/memory_modules_dimms.md
index 351c6fcd7..4f4d434fd 100644
--- a/collectors/proc.plugin/integrations/memory_modules_dimms.md
+++ b/src/collectors/proc.plugin/integrations/memory_modules_dimms.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/memory_modules_dimms.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/memory_modules_dimms.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Memory modules (DIMMs)"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Memory"
+learn_rel_path: "Collecting Metrics/Linux Systems/Memory"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -83,7 +83,7 @@ Metrics:
| Metric | Dimensions | Unit |
|:------|:----------|:----|
-| mem.edac_mc | correctable, uncorrectable, correctable_noinfo, uncorrectable_noinfo | errors/s |
+| mem.edac_mc_errors | correctable, uncorrectable, correctable_noinfo, uncorrectable_noinfo | errors |
### Per memory module
@@ -106,7 +106,7 @@ Metrics:
| Metric | Dimensions | Unit |
|:------|:----------|:----|
-| mem.edac_mc | correctable, uncorrectable | errors/s |
+| mem.edac_mc_errors | correctable, uncorrectable | errors |
@@ -117,10 +117,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ ecc_memory_mc_noinfo_correctable ](https://github.com/netdata/netdata/blob/master/health/health.d/memory.conf) | mem.edac_mc | memory controller ${label:controller} ECC correctable errors (unknown DIMM slot) in the last 10 minutes |
-| [ ecc_memory_mc_noinfo_uncorrectable ](https://github.com/netdata/netdata/blob/master/health/health.d/memory.conf) | mem.edac_mc | memory controller ${label:controller} ECC uncorrectable errors (unknown DIMM slot) in the last 10 minutes |
-| [ ecc_memory_dimm_correctable ](https://github.com/netdata/netdata/blob/master/health/health.d/memory.conf) | mem.edac_mc_dimm | DIMM ${label:dimm} controller ${label:controller} (location ${label:dimm_location}) ECC correctable errors in the last 10 minutes |
-| [ ecc_memory_dimm_uncorrectable ](https://github.com/netdata/netdata/blob/master/health/health.d/memory.conf) | mem.edac_mc_dimm | DIMM ${label:dimm} controller ${label:controller} (location ${label:dimm_location}) ECC uncorrectable errors in the last 10 minutes |
+| [ ecc_memory_mc_noinfo_correctable ](https://github.com/netdata/netdata/blob/master/src/health/health.d/memory.conf) | mem.edac_mc_errors | memory controller ${label:controller} ECC correctable errors (unknown DIMM slot) |
+| [ ecc_memory_mc_noinfo_uncorrectable ](https://github.com/netdata/netdata/blob/master/src/health/health.d/memory.conf) | mem.edac_mc_errors | memory controller ${label:controller} ECC uncorrectable errors (unknown DIMM slot) |
+| [ ecc_memory_dimm_correctable ](https://github.com/netdata/netdata/blob/master/src/health/health.d/memory.conf) | mem.edac_mc_dimm_errors | DIMM ${label:dimm} controller ${label:controller} (location ${label:dimm_location}) ECC correctable errors |
+| [ ecc_memory_dimm_uncorrectable ](https://github.com/netdata/netdata/blob/master/src/health/health.d/memory.conf) | mem.edac_mc_dimm_errors | DIMM ${label:dimm} controller ${label:controller} (location ${label:dimm_location}) ECC uncorrectable errors |
## Setup
diff --git a/collectors/proc.plugin/integrations/memory_statistics.md b/src/collectors/proc.plugin/integrations/memory_statistics.md
index 52f1bf530..a92df57a7 100644
--- a/collectors/proc.plugin/integrations/memory_statistics.md
+++ b/src/collectors/proc.plugin/integrations/memory_statistics.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/memory_statistics.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/memory_statistics.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Memory Statistics"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Memory"
+learn_rel_path: "Collecting Metrics/Linux Systems/Memory"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -111,8 +111,8 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 30min_ram_swapped_out ](https://github.com/netdata/netdata/blob/master/health/health.d/swap.conf) | mem.swapio | percentage of the system RAM swapped in the last 30 minutes |
-| [ oom_kill ](https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf) | mem.oom_kill | number of out of memory kills in the last 30 minutes |
+| [ 30min_ram_swapped_out ](https://github.com/netdata/netdata/blob/master/src/health/health.d/swap.conf) | mem.swapio | percentage of the system RAM swapped in the last 30 minutes |
+| [ oom_kill ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ram.conf) | mem.oom_kill | number of out of memory kills in the last 30 minutes |
## Setup
diff --git a/collectors/proc.plugin/integrations/memory_usage.md b/src/collectors/proc.plugin/integrations/memory_usage.md
index 141bd29ad..6c5168967 100644
--- a/collectors/proc.plugin/integrations/memory_usage.md
+++ b/src/collectors/proc.plugin/integrations/memory_usage.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/memory_usage.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/memory_usage.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Memory Usage"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Memory"
+learn_rel_path: "Collecting Metrics/Linux Systems/Memory"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -106,10 +106,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ ram_in_use ](https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf) | system.ram | system memory utilization |
-| [ ram_available ](https://github.com/netdata/netdata/blob/master/health/health.d/ram.conf) | mem.available | percentage of estimated amount of RAM available for userspace processes, without causing swapping |
-| [ used_swap ](https://github.com/netdata/netdata/blob/master/health/health.d/swap.conf) | mem.swap | swap memory utilization |
-| [ 1hour_memory_hw_corrupted ](https://github.com/netdata/netdata/blob/master/health/health.d/memory.conf) | mem.hwcorrupt | amount of memory corrupted due to a hardware failure |
+| [ ram_in_use ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ram.conf) | system.ram | system memory utilization |
+| [ ram_available ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ram.conf) | mem.available | percentage of estimated amount of RAM available for userspace processes, without causing swapping |
+| [ used_swap ](https://github.com/netdata/netdata/blob/master/src/health/health.d/swap.conf) | mem.swap | swap memory utilization |
+| [ 1hour_memory_hw_corrupted ](https://github.com/netdata/netdata/blob/master/src/health/health.d/memory.conf) | mem.hwcorrupt | amount of memory corrupted due to a hardware failure |
## Setup
diff --git a/collectors/proc.plugin/integrations/network_statistics.md b/src/collectors/proc.plugin/integrations/network_statistics.md
index 726fd9d61..460a8a60e 100644
--- a/collectors/proc.plugin/integrations/network_statistics.md
+++ b/src/collectors/proc.plugin/integrations/network_statistics.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/network_statistics.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/network_statistics.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Network statistics"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
+learn_rel_path: "Collecting Metrics/Linux Systems/Network"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -125,17 +125,17 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 1m_tcp_syn_queue_drops ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_listen.conf) | ip.tcp_syn_queue | average number of SYN requests was dropped due to the full TCP SYN queue over the last minute (SYN cookies were not enabled) |
-| [ 1m_tcp_syn_queue_cookies ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_listen.conf) | ip.tcp_syn_queue | average number of sent SYN cookies due to the full TCP SYN queue over the last minute |
-| [ 1m_tcp_accept_queue_overflows ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_listen.conf) | ip.tcp_accept_queue | average number of overflows in the TCP accept queue over the last minute |
-| [ 1m_tcp_accept_queue_drops ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_listen.conf) | ip.tcp_accept_queue | average number of dropped packets in the TCP accept queue over the last minute |
-| [ tcp_connections ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_conn.conf) | ip.tcpsock | TCP connections utilization |
-| [ 1m_ip_tcp_resets_sent ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf) | ip.tcphandshake | average number of sent TCP RESETS over the last minute |
-| [ 10s_ip_tcp_resets_sent ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf) | ip.tcphandshake | average number of sent TCP RESETS over the last 10 seconds. This can indicate a port scan, or that a service running on this host has crashed. Netdata will not send a clear notification for this alarm. |
-| [ 1m_ip_tcp_resets_received ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf) | ip.tcphandshake | average number of received TCP RESETS over the last minute |
-| [ 10s_ip_tcp_resets_received ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_resets.conf) | ip.tcphandshake | average number of received TCP RESETS over the last 10 seconds. This can be an indication that a service this host needs has crashed. Netdata will not send a clear notification for this alarm. |
-| [ 1m_ipv4_udp_receive_buffer_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/udp_errors.conf) | ipv4.udperrors | average number of UDP receive buffer errors over the last minute |
-| [ 1m_ipv4_udp_send_buffer_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/udp_errors.conf) | ipv4.udperrors | average number of UDP send buffer errors over the last minute |
+| [ 1m_tcp_syn_queue_drops ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_listen.conf) | ip.tcp_syn_queue | average number of SYN requests was dropped due to the full TCP SYN queue over the last minute (SYN cookies were not enabled) |
+| [ 1m_tcp_syn_queue_cookies ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_listen.conf) | ip.tcp_syn_queue | average number of sent SYN cookies due to the full TCP SYN queue over the last minute |
+| [ 1m_tcp_accept_queue_overflows ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_listen.conf) | ip.tcp_accept_queue | average number of overflows in the TCP accept queue over the last minute |
+| [ 1m_tcp_accept_queue_drops ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_listen.conf) | ip.tcp_accept_queue | average number of dropped packets in the TCP accept queue over the last minute |
+| [ tcp_connections ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_conn.conf) | ip.tcpsock | TCP connections utilization |
+| [ 1m_ip_tcp_resets_sent ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_resets.conf) | ip.tcphandshake | average number of sent TCP RESETS over the last minute |
+| [ 10s_ip_tcp_resets_sent ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_resets.conf) | ip.tcphandshake | average number of sent TCP RESETS over the last 10 seconds. This can indicate a port scan, or that a service running on this host has crashed. Netdata will not send a clear notification for this alarm. |
+| [ 1m_ip_tcp_resets_received ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_resets.conf) | ip.tcphandshake | average number of received TCP RESETS over the last minute |
+| [ 10s_ip_tcp_resets_received ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_resets.conf) | ip.tcphandshake | average number of received TCP RESETS over the last 10 seconds. This can be an indication that a service this host needs has crashed. Netdata will not send a clear notification for this alarm. |
+| [ 1m_ipv4_udp_receive_buffer_errors ](https://github.com/netdata/netdata/blob/master/src/health/health.d/udp_errors.conf) | ipv4.udperrors | average number of UDP receive buffer errors over the last minute |
+| [ 1m_ipv4_udp_send_buffer_errors ](https://github.com/netdata/netdata/blob/master/src/health/health.d/udp_errors.conf) | ipv4.udperrors | average number of UDP send buffer errors over the last minute |
## Setup
diff --git a/collectors/proc.plugin/integrations/nfs_client.md b/src/collectors/proc.plugin/integrations/nfs_client.md
index db5847714..3d3370b80 100644
--- a/collectors/proc.plugin/integrations/nfs_client.md
+++ b/src/collectors/proc.plugin/integrations/nfs_client.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/nfs_client.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/nfs_client.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "NFS Client"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Filesystem/NFS"
+learn_rel_path: "Collecting Metrics/Linux Systems/Filesystem/NFS"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/nfs_server.md b/src/collectors/proc.plugin/integrations/nfs_server.md
index 0c906b4d8..693e681c5 100644
--- a/collectors/proc.plugin/integrations/nfs_server.md
+++ b/src/collectors/proc.plugin/integrations/nfs_server.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/nfs_server.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/nfs_server.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "NFS Server"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Filesystem/NFS"
+learn_rel_path: "Collecting Metrics/Linux Systems/Filesystem/NFS"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/non-uniform_memory_access.md b/src/collectors/proc.plugin/integrations/non-uniform_memory_access.md
index 6f495fb79..3b55c65f1 100644
--- a/collectors/proc.plugin/integrations/non-uniform_memory_access.md
+++ b/src/collectors/proc.plugin/integrations/non-uniform_memory_access.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/non-uniform_memory_access.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/non-uniform_memory_access.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Non-Uniform Memory Access"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Memory"
+learn_rel_path: "Collecting Metrics/Linux Systems/Memory"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/page_types.md b/src/collectors/proc.plugin/integrations/page_types.md
index b228629b6..7dcb0f82d 100644
--- a/collectors/proc.plugin/integrations/page_types.md
+++ b/src/collectors/proc.plugin/integrations/page_types.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/page_types.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/page_types.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Page types"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Memory"
+learn_rel_path: "Collecting Metrics/Linux Systems/Memory"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/power_supply.md b/src/collectors/proc.plugin/integrations/power_supply.md
index 9a474e82a..53191dfff 100644
--- a/collectors/proc.plugin/integrations/power_supply.md
+++ b/src/collectors/proc.plugin/integrations/power_supply.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/power_supply.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/power_supply.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Power Supply"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Power Supply"
+learn_rel_path: "Collecting Metrics/Linux Systems/Power Supply"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -81,7 +81,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ linux_power_supply_capacity ](https://github.com/netdata/netdata/blob/master/health/health.d/linux_power_supply.conf) | powersupply.capacity | percentage of remaining power supply capacity |
+| [ linux_power_supply_capacity ](https://github.com/netdata/netdata/blob/master/src/health/health.d/linux_power_supply.conf) | powersupply.capacity | percentage of remaining power supply capacity |
## Setup
diff --git a/collectors/proc.plugin/integrations/pressure_stall_information.md b/src/collectors/proc.plugin/integrations/pressure_stall_information.md
index 53f4aa050..3af4da4a0 100644
--- a/collectors/proc.plugin/integrations/pressure_stall_information.md
+++ b/src/collectors/proc.plugin/integrations/pressure_stall_information.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/pressure_stall_information.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/pressure_stall_information.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Pressure Stall Information"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Pressure"
+learn_rel_path: "Collecting Metrics/Linux Systems/Pressure"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/sctp_statistics.md b/src/collectors/proc.plugin/integrations/sctp_statistics.md
index 15c0d424d..3c1cb7559 100644
--- a/collectors/proc.plugin/integrations/sctp_statistics.md
+++ b/src/collectors/proc.plugin/integrations/sctp_statistics.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/sctp_statistics.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/sctp_statistics.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "SCTP Statistics"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
+learn_rel_path: "Collecting Metrics/Linux Systems/Network"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/socket_statistics.md b/src/collectors/proc.plugin/integrations/socket_statistics.md
index d8ef26647..73fefc7c0 100644
--- a/collectors/proc.plugin/integrations/socket_statistics.md
+++ b/src/collectors/proc.plugin/integrations/socket_statistics.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/socket_statistics.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/socket_statistics.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Socket statistics"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
+learn_rel_path: "Collecting Metrics/Linux Systems/Network"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -82,8 +82,8 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ tcp_orphans ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_orphans.conf) | ipv4.sockstat_tcp_sockets | orphan IPv4 TCP sockets utilization |
-| [ tcp_memory ](https://github.com/netdata/netdata/blob/master/health/health.d/tcp_mem.conf) | ipv4.sockstat_tcp_mem | TCP memory utilization |
+| [ tcp_orphans ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_orphans.conf) | ipv4.sockstat_tcp_sockets | orphan IPv4 TCP sockets utilization |
+| [ tcp_memory ](https://github.com/netdata/netdata/blob/master/src/health/health.d/tcp_mem.conf) | ipv4.sockstat_tcp_mem | TCP memory utilization |
## Setup
diff --git a/collectors/proc.plugin/integrations/softirq_statistics.md b/src/collectors/proc.plugin/integrations/softirq_statistics.md
index f966cf971..2a130dcce 100644
--- a/collectors/proc.plugin/integrations/softirq_statistics.md
+++ b/src/collectors/proc.plugin/integrations/softirq_statistics.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/softirq_statistics.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/softirq_statistics.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "SoftIRQ statistics"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/CPU"
+learn_rel_path: "Collecting Metrics/Linux Systems/CPU"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/softnet_statistics.md b/src/collectors/proc.plugin/integrations/softnet_statistics.md
index 58e6cf6e5..fbbe08036 100644
--- a/collectors/proc.plugin/integrations/softnet_statistics.md
+++ b/src/collectors/proc.plugin/integrations/softnet_statistics.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/softnet_statistics.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/softnet_statistics.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Softnet Statistics"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
+learn_rel_path: "Collecting Metrics/Linux Systems/Network"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -108,8 +108,8 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 1min_netdev_backlog_exceeded ](https://github.com/netdata/netdata/blob/master/health/health.d/softnet.conf) | system.softnet_stat | average number of dropped packets in the last minute due to exceeded net.core.netdev_max_backlog |
-| [ 1min_netdev_budget_ran_outs ](https://github.com/netdata/netdata/blob/master/health/health.d/softnet.conf) | system.softnet_stat | average number of times ksoftirq ran out of sysctl net.core.netdev_budget or net.core.netdev_budget_usecs with work remaining over the last minute (this can be a cause for dropped packets) |
+| [ 1min_netdev_backlog_exceeded ](https://github.com/netdata/netdata/blob/master/src/health/health.d/softnet.conf) | system.softnet_stat | average number of dropped packets in the last minute due to exceeded net.core.netdev_max_backlog |
+| [ 1min_netdev_budget_ran_outs ](https://github.com/netdata/netdata/blob/master/src/health/health.d/softnet.conf) | system.softnet_stat | average number of times ksoftirq ran out of sysctl net.core.netdev_budget or net.core.netdev_budget_usecs with work remaining over the last minute (this can be a cause for dropped packets) |
## Setup
diff --git a/collectors/proc.plugin/integrations/synproxy.md b/src/collectors/proc.plugin/integrations/synproxy.md
index 2db17ef6f..a63e6cbc0 100644
--- a/collectors/proc.plugin/integrations/synproxy.md
+++ b/src/collectors/proc.plugin/integrations/synproxy.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/synproxy.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/synproxy.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Synproxy"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Firewall"
+learn_rel_path: "Collecting Metrics/Linux Systems/Firewall"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/system_load_average.md b/src/collectors/proc.plugin/integrations/system_load_average.md
index 6e986d90c..51f4f14ba 100644
--- a/collectors/proc.plugin/integrations/system_load_average.md
+++ b/src/collectors/proc.plugin/integrations/system_load_average.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/system_load_average.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/system_load_average.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "System Load Average"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/System"
+learn_rel_path: "Collecting Metrics/Linux Systems/System"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -98,11 +98,11 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ load_cpu_number ](https://github.com/netdata/netdata/blob/master/health/health.d/load.conf) | system.load | number of active CPU cores in the system |
-| [ load_average_15 ](https://github.com/netdata/netdata/blob/master/health/health.d/load.conf) | system.load | system fifteen-minute load average |
-| [ load_average_5 ](https://github.com/netdata/netdata/blob/master/health/health.d/load.conf) | system.load | system five-minute load average |
-| [ load_average_1 ](https://github.com/netdata/netdata/blob/master/health/health.d/load.conf) | system.load | system one-minute load average |
-| [ active_processes ](https://github.com/netdata/netdata/blob/master/health/health.d/processes.conf) | system.active_processes | system process IDs (PID) space utilization |
+| [ load_cpu_number ](https://github.com/netdata/netdata/blob/master/src/health/health.d/load.conf) | system.load | number of active CPU cores in the system |
+| [ load_average_15 ](https://github.com/netdata/netdata/blob/master/src/health/health.d/load.conf) | system.load | system fifteen-minute load average |
+| [ load_average_5 ](https://github.com/netdata/netdata/blob/master/src/health/health.d/load.conf) | system.load | system five-minute load average |
+| [ load_average_1 ](https://github.com/netdata/netdata/blob/master/src/health/health.d/load.conf) | system.load | system one-minute load average |
+| [ active_processes ](https://github.com/netdata/netdata/blob/master/src/health/health.d/processes.conf) | system.active_processes | system process IDs (PID) space utilization |
## Setup
diff --git a/collectors/proc.plugin/integrations/system_statistics.md b/src/collectors/proc.plugin/integrations/system_statistics.md
index f3df1a19a..264e40bd5 100644
--- a/collectors/proc.plugin/integrations/system_statistics.md
+++ b/src/collectors/proc.plugin/integrations/system_statistics.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/system_statistics.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/system_statistics.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "System statistics"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/System"
+learn_rel_path: "Collecting Metrics/Linux Systems/System"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -122,9 +122,9 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ 10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf) | system.cpu | average CPU utilization over the last 10 minutes (excluding iowait, nice and steal) |
-| [ 10min_cpu_iowait ](https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf) | system.cpu | average CPU iowait time over the last 10 minutes |
-| [ 20min_steal_cpu ](https://github.com/netdata/netdata/blob/master/health/health.d/cpu.conf) | system.cpu | average CPU steal time over the last 20 minutes |
+| [ 10min_cpu_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cpu.conf) | system.cpu | average CPU utilization over the last 10 minutes (excluding iowait, nice and steal) |
+| [ 10min_cpu_iowait ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cpu.conf) | system.cpu | average CPU iowait time over the last 10 minutes |
+| [ 20min_steal_cpu ](https://github.com/netdata/netdata/blob/master/src/health/health.d/cpu.conf) | system.cpu | average CPU steal time over the last 20 minutes |
## Setup
@@ -151,7 +151,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/proc.plugin/integrations/system_uptime.md b/src/collectors/proc.plugin/integrations/system_uptime.md
index 0954c0642..4b7e21188 100644
--- a/collectors/proc.plugin/integrations/system_uptime.md
+++ b/src/collectors/proc.plugin/integrations/system_uptime.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/system_uptime.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/system_uptime.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "System Uptime"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/System"
+learn_rel_path: "Collecting Metrics/Linux Systems/System"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/wireless_network_interfaces.md b/src/collectors/proc.plugin/integrations/wireless_network_interfaces.md
index a8d2406ee..4288a1ebd 100644
--- a/collectors/proc.plugin/integrations/wireless_network_interfaces.md
+++ b/src/collectors/proc.plugin/integrations/wireless_network_interfaces.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/wireless_network_interfaces.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/wireless_network_interfaces.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "Wireless network interfaces"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
+learn_rel_path: "Collecting Metrics/Linux Systems/Network"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/integrations/zfs_adaptive_replacement_cache.md b/src/collectors/proc.plugin/integrations/zfs_adaptive_replacement_cache.md
index c200ba673..d2d9378f4 100644
--- a/collectors/proc.plugin/integrations/zfs_adaptive_replacement_cache.md
+++ b/src/collectors/proc.plugin/integrations/zfs_adaptive_replacement_cache.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/zfs_adaptive_replacement_cache.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/zfs_adaptive_replacement_cache.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "ZFS Adaptive Replacement Cache"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Filesystem/ZFS"
+learn_rel_path: "Collecting Metrics/Linux Systems/Filesystem/ZFS"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -99,7 +99,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ zfs_memory_throttle ](https://github.com/netdata/netdata/blob/master/health/health.d/zfs.conf) | zfs.memory_ops | number of times ZFS had to limit the ARC growth in the last 10 minutes |
+| [ zfs_memory_throttle ](https://github.com/netdata/netdata/blob/master/src/health/health.d/zfs.conf) | zfs.memory_ops | number of times ZFS had to limit the ARC growth in the last 10 minutes |
## Setup
diff --git a/collectors/proc.plugin/integrations/zfs_pools.md b/src/collectors/proc.plugin/integrations/zfs_pools.md
index 2985d39b0..f18c82baf 100644
--- a/collectors/proc.plugin/integrations/zfs_pools.md
+++ b/src/collectors/proc.plugin/integrations/zfs_pools.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/zfs_pools.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/zfs_pools.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "ZFS Pools"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Filesystem/ZFS"
+learn_rel_path: "Collecting Metrics/Linux Systems/Filesystem/ZFS"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -78,8 +78,8 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ zfs_pool_state_warn ](https://github.com/netdata/netdata/blob/master/health/health.d/zfs.conf) | zfspool.state | ZFS pool ${label:pool} state is degraded |
-| [ zfs_pool_state_crit ](https://github.com/netdata/netdata/blob/master/health/health.d/zfs.conf) | zfspool.state | ZFS pool ${label:pool} state is faulted or unavail |
+| [ zfs_pool_state_warn ](https://github.com/netdata/netdata/blob/master/src/health/health.d/zfs.conf) | zfspool.state | ZFS pool ${label:pool} state is degraded |
+| [ zfs_pool_state_crit ](https://github.com/netdata/netdata/blob/master/src/health/health.d/zfs.conf) | zfspool.state | ZFS pool ${label:pool} state is faulted or unavail |
## Setup
diff --git a/collectors/proc.plugin/integrations/zram.md b/src/collectors/proc.plugin/integrations/zram.md
index 111b17c62..b80a72ab1 100644
--- a/collectors/proc.plugin/integrations/zram.md
+++ b/src/collectors/proc.plugin/integrations/zram.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/integrations/zram.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/proc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/integrations/zram.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/proc.plugin/metadata.yaml"
sidebar_label: "ZRAM"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Memory"
+learn_rel_path: "Collecting Metrics/Linux Systems/Memory"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/proc.plugin/ipc.c b/src/collectors/proc.plugin/ipc.c
index 204977bdf..6d7d920f0 100644
--- a/collectors/proc.plugin/ipc.c
+++ b/src/collectors/proc.plugin/ipc.c
@@ -352,8 +352,8 @@ int do_ipc(int update_every, usec_t dt) {
}
// variables
- semaphores_max = rrdvar_custom_host_variable_add_and_acquire(localhost, "ipc_semaphores_max");
- arrays_max = rrdvar_custom_host_variable_add_and_acquire(localhost, "ipc_semaphores_arrays_max");
+ semaphores_max = rrdvar_host_variable_add_and_acquire(localhost, "ipc_semaphores_max");
+ arrays_max = rrdvar_host_variable_add_and_acquire(localhost, "ipc_semaphores_arrays_max");
}
struct stat stbuf;
@@ -373,8 +373,10 @@ int do_ipc(int update_every, usec_t dt) {
collector_error("Unable to fetch semaphore limits.");
}
else {
- if(semaphores_max) rrdvar_custom_host_variable_set(localhost, semaphores_max, limits.semmns);
- if(arrays_max) rrdvar_custom_host_variable_set(localhost, arrays_max, limits.semmni);
+ if(semaphores_max)
+ rrdvar_host_variable_set(localhost, semaphores_max, limits.semmns);
+ if(arrays_max)
+ rrdvar_host_variable_set(localhost, arrays_max, limits.semmni);
st_arrays->red = limits.semmni;
st_semaphores->red = limits.semmns;
diff --git a/collectors/proc.plugin/plugin_proc.c b/src/collectors/proc.plugin/plugin_proc.c
index 3f11aaf6c..7742b344f 100644
--- a/collectors/proc.plugin/plugin_proc.c
+++ b/src/collectors/proc.plugin/plugin_proc.c
@@ -157,59 +157,57 @@ void *proc_main(void *ptr)
netdev_thread, THREAD_NETDEV_NAME, NETDATA_THREAD_OPTION_JOINABLE, netdev_main, netdev_thread);
}
- netdata_thread_cleanup_push(proc_main_cleanup, ptr);
+ netdata_thread_cleanup_push(proc_main_cleanup, ptr)
+ {
+ config_get_boolean("plugin:proc", "/proc/pagetypeinfo", CONFIG_BOOLEAN_NO);
- {
- config_get_boolean("plugin:proc", "/proc/pagetypeinfo", CONFIG_BOOLEAN_NO);
+ // check the enabled status for each module
+ int i;
+ for(i = 0; proc_modules[i].name; i++) {
+ struct proc_module *pm = &proc_modules[i];
- // check the enabled status for each module
- int i;
- for(i = 0; proc_modules[i].name; i++) {
- struct proc_module *pm = &proc_modules[i];
+ pm->enabled = config_get_boolean("plugin:proc", pm->name, CONFIG_BOOLEAN_YES);
+ pm->rd = NULL;
- pm->enabled = config_get_boolean("plugin:proc", pm->name, CONFIG_BOOLEAN_YES);
- pm->rd = NULL;
-
- worker_register_job_name(i, proc_modules[i].dim);
- }
+ worker_register_job_name(i, proc_modules[i].dim);
+ }
- usec_t step = localhost->rrd_update_every * USEC_PER_SEC;
- heartbeat_t hb;
- heartbeat_init(&hb);
+ usec_t step = localhost->rrd_update_every * USEC_PER_SEC;
+ heartbeat_t hb;
+ heartbeat_init(&hb);
- inside_lxc_container = is_lxcfs_proc_mounted();
+ inside_lxc_container = is_lxcfs_proc_mounted();
#define LGS_MODULE_ID 0
- ND_LOG_STACK lgs[] = {
- [LGS_MODULE_ID] = ND_LOG_FIELD_TXT(NDF_MODULE, "proc.plugin"),
- ND_LOG_FIELD_END(),
- };
- ND_LOG_STACK_PUSH(lgs);
-
- while(service_running(SERVICE_COLLECTORS)) {
- worker_is_idle();
- usec_t hb_dt = heartbeat_next(&hb, step);
-
- if(unlikely(!service_running(SERVICE_COLLECTORS)))
- break;
-
- for(i = 0; proc_modules[i].name; i++) {
- if(unlikely(!service_running(SERVICE_COLLECTORS)))
- break;
-
- struct proc_module *pm = &proc_modules[i];
- if(unlikely(!pm->enabled))
- continue;
-
- worker_is_busy(i);
- lgs[LGS_MODULE_ID] = ND_LOG_FIELD_CB(NDF_MODULE, log_proc_module, pm);
- pm->enabled = !pm->func(localhost->rrd_update_every, hb_dt);
- lgs[LGS_MODULE_ID] = ND_LOG_FIELD_TXT(NDF_MODULE, "proc.plugin");
- }
- }
- }
+ ND_LOG_STACK lgs[] = {
+ [LGS_MODULE_ID] = ND_LOG_FIELD_TXT(NDF_MODULE, "proc.plugin"),
+ ND_LOG_FIELD_END(),
+ };
+ ND_LOG_STACK_PUSH(lgs);
+
+ while(service_running(SERVICE_COLLECTORS)) {
+ worker_is_idle();
+ usec_t hb_dt = heartbeat_next(&hb, step);
+
+ if(unlikely(!service_running(SERVICE_COLLECTORS)))
+ break;
+ for(i = 0; proc_modules[i].name; i++) {
+ if(unlikely(!service_running(SERVICE_COLLECTORS)))
+ break;
+
+ struct proc_module *pm = &proc_modules[i];
+ if(unlikely(!pm->enabled))
+ continue;
+
+ worker_is_busy(i);
+ lgs[LGS_MODULE_ID] = ND_LOG_FIELD_CB(NDF_MODULE, log_proc_module, pm);
+ pm->enabled = !pm->func(localhost->rrd_update_every, hb_dt);
+ lgs[LGS_MODULE_ID] = ND_LOG_FIELD_TXT(NDF_MODULE, "proc.plugin");
+ }
+ }
+ }
netdata_thread_cleanup_pop(1);
return NULL;
}
diff --git a/collectors/proc.plugin/plugin_proc.h b/src/collectors/proc.plugin/plugin_proc.h
index e4fc105ba..187e76a97 100644
--- a/collectors/proc.plugin/plugin_proc.h
+++ b/src/collectors/proc.plugin/plugin_proc.h
@@ -54,7 +54,7 @@ extern unsigned long long zfs_arcstats_shrinkable_cache_size_bytes;
extern bool inside_lxc_container;
// netdev renames
-void netdev_rename_device_add(
+void cgroup_rename_task_add(
const char *host_device,
const char *container_device,
const char *container_name,
@@ -62,7 +62,7 @@ void netdev_rename_device_add(
const char *ctx_prefix,
const DICTIONARY_ITEM *cgroup_netdev_link);
-void netdev_rename_device_del(const char *host_device);
+void cgroup_rename_task_device_del(const char *host_device);
#include "proc_self_mountinfo.h"
#include "proc_pressure.h"
diff --git a/collectors/proc.plugin/proc_diskstats.c b/src/collectors/proc.plugin/proc_diskstats.c
index 475d90835..4ff617ff9 100644
--- a/collectors/proc.plugin/proc_diskstats.c
+++ b/src/collectors/proc.plugin/proc_diskstats.c
@@ -213,7 +213,7 @@ static SIMPLE_PATTERN *excluded_disks = NULL;
static unsigned long long int bcache_read_number_with_units(const char *filename) {
char buffer[50 + 1];
- if(read_file(filename, buffer, 50) == 0) {
+ if(read_txt_file(filename, buffer, sizeof(buffer)) == 0) {
static int unknown_units_error = 10;
char *end = NULL;
@@ -547,9 +547,9 @@ static inline char *get_disk_model(char *device) {
char buffer[256 + 1];
snprintfz(path, sizeof(path) - 1, "%s/%s/device/model", path_to_sys_block, device);
- if(read_file(path, buffer, 256) != 0) {
+ if(read_txt_file(path, buffer, sizeof(buffer)) != 0) {
snprintfz(path, sizeof(path) - 1, "%s/%s/device/name", path_to_sys_block, device);
- if(read_file(path, buffer, 256) != 0)
+ if(read_txt_file(path, buffer, sizeof(buffer)) != 0)
return NULL;
}
@@ -565,7 +565,7 @@ static inline char *get_disk_serial(char *device) {
char buffer[256 + 1];
snprintfz(path, sizeof(path) - 1, "%s/%s/device/serial", path_to_sys_block, device);
- if(read_file(path, buffer, 256) != 0)
+ if(read_txt_file(path, buffer, sizeof(buffer)) != 0)
return NULL;
return strdupz(buffer);
@@ -778,7 +778,7 @@ static struct disk *get_disk(unsigned long major, unsigned long minor, char *dis
strncat(uuid_filename, "/dm/uuid", FILENAME_MAX - size);
char device_uuid[RRD_ID_LENGTH_MAX + 1];
- if (!read_file(uuid_filename, device_uuid, RRD_ID_LENGTH_MAX) && !strncmp(device_uuid, "LVM-", 4)) {
+ if (!read_txt_file(uuid_filename, device_uuid, sizeof(device_uuid)) && !strncmp(device_uuid, "LVM-", 4)) {
trim(device_uuid);
char chart_id[RRD_ID_LENGTH_MAX + 1];
@@ -1033,13 +1033,7 @@ static void add_labels_to_disk(struct disk *d, RRDSET *st) {
rrdlabels_add(st->rrdlabels, "device_type", get_disk_type_string(d->type), RRDLABEL_SRC_AUTO);
}
-static int diskstats_function_block_devices(BUFFER *wb, int timeout __maybe_unused, const char *function __maybe_unused,
- void *collector_data __maybe_unused,
- rrd_function_result_callback_t result_cb, void *result_cb_data,
- rrd_function_is_cancelled_cb_t is_cancelled_cb, void *is_cancelled_cb_data,
- rrd_function_register_canceller_cb_t register_canceller_cb __maybe_unused,
- void *register_canceller_cb_data __maybe_unused) {
-
+static int diskstats_function_block_devices(BUFFER *wb, const char *function __maybe_unused) {
buffer_flush(wb);
wb->content_type = CT_APPLICATION_JSON;
buffer_json_initialize(wb, "\"", "\"", 0, true, BUFFER_JSON_OPTIONS_DEFAULT);
@@ -1048,6 +1042,7 @@ static int diskstats_function_block_devices(BUFFER *wb, int timeout __maybe_unus
buffer_json_member_add_uint64(wb, "status", HTTP_RESP_OK);
buffer_json_member_add_string(wb, "type", "table");
buffer_json_member_add_time_t(wb, "update_every", 1);
+ buffer_json_member_add_boolean(wb, "has_history", false);
buffer_json_member_add_string(wb, "help", RRDFUNCTIONS_DISKSTATS_HELP);
buffer_json_member_add_array(wb, "data");
@@ -1318,16 +1313,7 @@ static int diskstats_function_block_devices(BUFFER *wb, int timeout __maybe_unus
buffer_json_member_add_time_t(wb, "expires", now_realtime_sec() + 1);
buffer_json_finalize(wb);
- int response = HTTP_RESP_OK;
- if(is_cancelled_cb && is_cancelled_cb(is_cancelled_cb_data)) {
- buffer_flush(wb);
- response = HTTP_RESP_CLIENT_CLOSED_REQUEST;
- }
-
- if(result_cb)
- result_cb(wb, response, result_cb_data);
-
- return response;
+ return HTTP_RESP_OK;
}
static void diskstats_cleanup_disks() {
@@ -1474,6 +1460,11 @@ int do_proc_diskstats(int update_every, usec_t dt) {
excluded_disks = simple_pattern_create(
config_get(CONFIG_SECTION_PLUGIN_PROC_DISKSTATS, "exclude disks", DEFAULT_EXCLUDED_DISKS), NULL,
SIMPLE_PATTERN_EXACT, true);
+
+ rrd_function_add_inline(localhost, NULL, "block-devices", 10,
+ RRDFUNCTIONS_PRIORITY_DEFAULT, RRDFUNCTIONS_DISKSTATS_HELP,
+ "top", HTTP_ACCESS_ANONYMOUS_DATA,
+ diskstats_function_block_devices);
}
// --------------------------------------------------------------------------
@@ -1488,12 +1479,6 @@ int do_proc_diskstats(int update_every, usec_t dt) {
ff = procfile_readall(ff);
if(unlikely(!ff)) return 0; // we return 0, so that we will retry to open it next time
- static bool add_func = true;
- if (add_func) {
- rrd_function_add(localhost, NULL, "block-devices", 10, RRDFUNCTIONS_DISKSTATS_HELP, true, diskstats_function_block_devices, NULL);
- add_func = false;
- }
-
size_t lines = procfile_lines(ff), l;
collected_number system_read_kb = 0, system_write_kb = 0;
diff --git a/collectors/proc.plugin/proc_interrupts.c b/src/collectors/proc.plugin/proc_interrupts.c
index 37071b22f..aa9bd0eb5 100644
--- a/collectors/proc.plugin/proc_interrupts.c
+++ b/src/collectors/proc.plugin/proc_interrupts.c
@@ -60,7 +60,7 @@ int do_proc_interrupts(int update_every, usec_t dt) {
struct interrupt *irrs = NULL;
if(unlikely(do_per_core == CONFIG_BOOLEAN_INVALID))
- do_per_core = config_get_boolean_ondemand(CONFIG_SECTION_PLUGIN_PROC_INTERRUPTS, "interrupts per core", CONFIG_BOOLEAN_AUTO);
+ do_per_core = config_get_boolean_ondemand(CONFIG_SECTION_PLUGIN_PROC_INTERRUPTS, "interrupts per core", CONFIG_BOOLEAN_NO);
if(unlikely(!ff)) {
char filename[FILENAME_MAX + 1];
diff --git a/collectors/proc.plugin/proc_loadavg.c b/src/collectors/proc.plugin/proc_loadavg.c
index 106cf9087..c9339525e 100644
--- a/collectors/proc.plugin/proc_loadavg.c
+++ b/src/collectors/proc.plugin/proc_loadavg.c
@@ -95,7 +95,7 @@ int do_proc_loadavg(int update_every, usec_t dt) {
if(likely(do_all_processes)) {
static RRDSET *processes_chart = NULL;
static RRDDIM *rd_active = NULL;
- static const RRDSETVAR_ACQUIRED *rd_pidmax;
+ static const RRDVAR_ACQUIRED *rd_pidmax;
if(unlikely(!processes_chart)) {
processes_chart = rrdset_create_localhost(
@@ -114,11 +114,11 @@ int do_proc_loadavg(int update_every, usec_t dt) {
);
rd_active = rrddim_add(processes_chart, "active", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
- rd_pidmax = rrdsetvar_custom_chart_variable_add_and_acquire(processes_chart, "pidmax");
+ rd_pidmax = rrdvar_chart_variable_add_and_acquire(processes_chart, "pidmax");
}
rrddim_set_by_pointer(processes_chart, rd_active, active_processes);
- rrdsetvar_custom_chart_variable_set(processes_chart, rd_pidmax, max_processes);
+ rrdvar_chart_variable_set(processes_chart, rd_pidmax, max_processes);
rrdset_done(processes_chart);
}
diff --git a/collectors/proc.plugin/proc_mdstat.c b/src/collectors/proc.plugin/proc_mdstat.c
index 3857d9ec4..3857d9ec4 100644
--- a/collectors/proc.plugin/proc_mdstat.c
+++ b/src/collectors/proc.plugin/proc_mdstat.c
diff --git a/collectors/proc.plugin/proc_meminfo.c b/src/collectors/proc.plugin/proc_meminfo.c
index a357cc782..a357cc782 100644
--- a/collectors/proc.plugin/proc_meminfo.c
+++ b/src/collectors/proc.plugin/proc_meminfo.c
diff --git a/collectors/proc.plugin/proc_net_dev.c b/src/collectors/proc.plugin/proc_net_dev.c
index b39f39683..d29bb7a72 100644
--- a/collectors/proc.plugin/proc_net_dev.c
+++ b/src/collectors/proc.plugin/proc_net_dev.c
@@ -1,6 +1,7 @@
// SPDX-License-Identifier: GPL-3.0-or-later
#include "plugin_proc.h"
+#include "proc_net_dev_renames.h"
#define PLUGIN_PROC_MODULE_NETDEV_NAME "/proc/net/dev"
#define CONFIG_SECTION_PLUGIN_PROC_NETDEV "plugin:" PLUGIN_PROC_CONFIG_NAME ":" PLUGIN_PROC_MODULE_NETDEV_NAME
@@ -11,10 +12,7 @@
#define READ_RETRY_PERIOD 60 // seconds
-void cgroup_netdev_reset_all(void);
-void cgroup_netdev_release(const DICTIONARY_ITEM *link);
-const void *cgroup_netdev_dup(const DICTIONARY_ITEM *link);
-void cgroup_netdev_add_bandwidth(const DICTIONARY_ITEM *link, NETDATA_DOUBLE received, NETDATA_DOUBLE sent);
+time_t double_linked_device_collect_delay_secs = 120;
enum {
NETDEV_DUPLEX_UNKNOWN,
@@ -22,8 +20,7 @@ enum {
NETDEV_DUPLEX_FULL
};
-static const char *get_duplex_string(int duplex)
-{
+static const char *get_duplex_string(int duplex) {
switch (duplex) {
case NETDEV_DUPLEX_FULL:
return "full";
@@ -44,8 +41,7 @@ enum {
NETDEV_OPERSTATE_UP
};
-static inline int get_operstate(char *operstate)
-{
+static inline int get_operstate(char *operstate) {
// As defined in https://www.kernel.org/doc/Documentation/ABI/testing/sysfs-class-net
if (!strcmp(operstate, "up"))
return NETDEV_OPERSTATE_UP;
@@ -63,8 +59,7 @@ static inline int get_operstate(char *operstate)
return NETDEV_OPERSTATE_UNKNOWN;
}
-static const char *get_operstate_string(int operstate)
-{
+static const char *get_operstate_string(int operstate) {
switch (operstate) {
case NETDEV_OPERSTATE_UP:
return "up";
@@ -92,12 +87,12 @@ static struct netdev {
size_t len;
// flags
- int virtual;
- int configured;
+ bool virtual;
+ bool configured;
int enabled;
- int updated;
-
+ bool updated;
bool function_ready;
+ bool double_linked; // iflink != ifindex
time_t discover_time;
@@ -242,7 +237,7 @@ static struct netdev {
RRDDIM *rd_mtu;
char *filename_speed;
- const RRDSETVAR_ACQUIRED *chart_var_speed;
+ const RRDVAR_ACQUIRED *chart_var_speed;
char *filename_duplex;
char *filename_operstate;
@@ -251,10 +246,8 @@ static struct netdev {
const DICTIONARY_ITEM *cgroup_netdev_link;
- struct netdev *next;
-} *netdev_root = NULL, *netdev_last_used = NULL;
-
-static size_t netdev_added = 0, netdev_found = 0;
+ struct netdev *prev, *next;
+} *netdev_root = NULL;
// ----------------------------------------------------------------------------
@@ -378,133 +371,21 @@ static void netdev_free(struct netdev *d) {
freez((void *)d->filename_carrier);
freez((void *)d->filename_mtu);
freez((void *)d);
- netdev_added--;
}
-// ----------------------------------------------------------------------------
-// netdev renames
-
-static struct netdev_rename {
- const char *host_device;
- uint32_t hash;
-
- const char *container_device;
- const char *container_name;
- const char *ctx_prefix;
-
- RRDLABELS *chart_labels;
-
- int processed;
-
- const DICTIONARY_ITEM *cgroup_netdev_link;
-
- struct netdev_rename *next;
-} *netdev_rename_root = NULL;
-
-static int netdev_pending_renames = 0;
-static netdata_mutex_t netdev_rename_mutex = NETDATA_MUTEX_INITIALIZER;
-static netdata_mutex_t netdev_dev_mutex = NETDATA_MUTEX_INITIALIZER;
-
-static struct netdev_rename *netdev_rename_find(const char *host_device, uint32_t hash) {
- struct netdev_rename *r;
+static netdata_mutex_t netdev_mutex = NETDATA_MUTEX_INITIALIZER;
- for(r = netdev_rename_root; r ; r = r->next)
- if(r->hash == hash && !strcmp(host_device, r->host_device))
- return r;
-
- return NULL;
-}
-
-// other threads can call this function to register a rename to a netdev
-void netdev_rename_device_add(
- const char *host_device,
- const char *container_device,
- const char *container_name,
- RRDLABELS *labels,
- const char *ctx_prefix,
- const DICTIONARY_ITEM *cgroup_netdev_link)
-{
- netdata_mutex_lock(&netdev_rename_mutex);
-
- uint32_t hash = simple_hash(host_device);
- struct netdev_rename *r = netdev_rename_find(host_device, hash);
- if(!r) {
- r = callocz(1, sizeof(struct netdev_rename));
- r->host_device = strdupz(host_device);
- r->container_device = strdupz(container_device);
- r->container_name = strdupz(container_name);
- r->ctx_prefix = strdupz(ctx_prefix);
- r->chart_labels = rrdlabels_create();
- rrdlabels_migrate_to_these(r->chart_labels, labels);
- r->hash = hash;
- r->next = netdev_rename_root;
- r->processed = 0;
- r->cgroup_netdev_link = cgroup_netdev_link;
-
- netdev_rename_root = r;
- netdev_pending_renames++;
- collector_info("CGROUP: registered network interface rename for '%s' as '%s' under '%s'", r->host_device, r->container_device, r->container_name);
- }
- else {
- if(strcmp(r->container_device, container_device) != 0 || strcmp(r->container_name, container_name) != 0) {
- freez((void *) r->container_device);
- freez((void *) r->container_name);
-
- r->container_device = strdupz(container_device);
- r->container_name = strdupz(container_name);
-
- rrdlabels_migrate_to_these(r->chart_labels, labels);
-
- r->processed = 0;
- r->cgroup_netdev_link = cgroup_netdev_link;
-
- netdev_pending_renames++;
- collector_info("CGROUP: altered network interface rename for '%s' as '%s' under '%s'", r->host_device, r->container_device, r->container_name);
- }
- }
-
- netdata_mutex_unlock(&netdev_rename_mutex);
-}
-
-// other threads can call this function to delete a rename to a netdev
-void netdev_rename_device_del(const char *host_device) {
- netdata_mutex_lock(&netdev_rename_mutex);
-
- struct netdev_rename *r, *last = NULL;
-
- uint32_t hash = simple_hash(host_device);
- for(r = netdev_rename_root; r ; last = r, r = r->next) {
- if (r->hash == hash && !strcmp(host_device, r->host_device)) {
- if (netdev_rename_root == r)
- netdev_rename_root = r->next;
- else if (last)
- last->next = r->next;
-
- if(!r->processed)
- netdev_pending_renames--;
-
- collector_info("CGROUP: unregistered network interface rename for '%s' as '%s' under '%s'", r->host_device, r->container_device, r->container_name);
-
- freez((void *) r->host_device);
- freez((void *) r->container_name);
- freez((void *) r->container_device);
- freez((void *) r->ctx_prefix);
- rrdlabels_destroy(r->chart_labels);
- cgroup_netdev_release(r->cgroup_netdev_link);
- freez((void *) r);
- break;
- }
- }
-
- netdata_mutex_unlock(&netdev_rename_mutex);
-}
+// ----------------------------------------------------------------------------
-static inline void netdev_rename_cgroup(struct netdev *d, struct netdev_rename *r) {
- collector_info("CGROUP: renaming network interface '%s' as '%s' under '%s'", r->host_device, r->container_device, r->container_name);
+static inline void netdev_rename(struct netdev *d, struct rename_task *r) {
+ collector_info("CGROUP: renaming network interface '%s' as '%s' under '%s'", d->name, r->container_device, r->container_name);
netdev_charts_release(d);
netdev_free_chart_strings(d);
+
+ cgroup_netdev_release(d->cgroup_netdev_link);
d->cgroup_netdev_link = cgroup_netdev_dup(r->cgroup_netdev_link);
+ d->discover_time = 0;
char buffer[RRD_ID_LENGTH_MAX + 1];
@@ -581,42 +462,18 @@ static inline void netdev_rename_cgroup(struct netdev *d, struct netdev_rename *
d->flipped = 1;
}
-static inline void netdev_rename(struct netdev *d) {
- struct netdev_rename *r = netdev_rename_find(d->name, d->hash);
- if(unlikely(r && !r->processed)) {
- netdev_rename_cgroup(d, r);
- r->processed = 1;
- d->discover_time = 0;
- netdev_pending_renames--;
+static void netdev_rename_this_device(struct netdev *d) {
+ const DICTIONARY_ITEM *item = dictionary_get_and_acquire_item(netdev_renames, d->name);
+ if(item) {
+ struct rename_task *r = dictionary_acquired_item_value(item);
+ netdev_rename(d, r);
+ dictionary_acquired_item_release(netdev_renames, item);
}
}
-static inline void netdev_rename_lock(struct netdev *d) {
- netdata_mutex_lock(&netdev_rename_mutex);
- netdev_rename(d);
- netdata_mutex_unlock(&netdev_rename_mutex);
-}
-
-static inline void netdev_rename_all_lock(void) {
- netdata_mutex_lock(&netdev_rename_mutex);
-
- struct netdev *d;
- for(d = netdev_root; d ; d = d->next)
- netdev_rename(d);
-
- netdev_pending_renames = 0;
- netdata_mutex_unlock(&netdev_rename_mutex);
-}
-
// ----------------------------------------------------------------------------
-int netdev_function_net_interfaces(BUFFER *wb, int timeout __maybe_unused, const char *function __maybe_unused,
- void *collector_data __maybe_unused,
- rrd_function_result_callback_t result_cb, void *result_cb_data,
- rrd_function_is_cancelled_cb_t is_cancelled_cb, void *is_cancelled_cb_data,
- rrd_function_register_canceller_cb_t register_canceller_cb __maybe_unused,
- void *register_canceller_cb_data __maybe_unused) {
-
+int netdev_function_net_interfaces(BUFFER *wb, const char *function __maybe_unused) {
buffer_flush(wb);
wb->content_type = CT_APPLICATION_JSON;
buffer_json_initialize(wb, "\"", "\"", 0, true, BUFFER_JSON_OPTIONS_DEFAULT);
@@ -625,6 +482,7 @@ int netdev_function_net_interfaces(BUFFER *wb, int timeout __maybe_unused, const
buffer_json_member_add_uint64(wb, "status", HTTP_RESP_OK);
buffer_json_member_add_string(wb, "type", "table");
buffer_json_member_add_time_t(wb, "update_every", 1);
+ buffer_json_member_add_boolean(wb, "has_history", false);
buffer_json_member_add_string(wb, "help", RRDFUNCTIONS_NETDEV_HELP);
buffer_json_member_add_array(wb, "data");
@@ -637,11 +495,11 @@ int netdev_function_net_interfaces(BUFFER *wb, int timeout __maybe_unused, const
double max_drops_rx = 0.0;
double max_drops_tx = 0.0;
- netdata_mutex_lock(&netdev_dev_mutex);
+ netdata_mutex_lock(&netdev_mutex);
RRDDIM *rd = NULL;
- for (struct netdev *d = netdev_root; d != netdev_last_used; d = d->next) {
+ for (struct netdev *d = netdev_root; d ; d = d->next) {
if (unlikely(!d->function_ready))
continue;
@@ -701,7 +559,7 @@ int netdev_function_net_interfaces(BUFFER *wb, int timeout __maybe_unused, const
buffer_json_array_close(wb);
}
- netdata_mutex_unlock(&netdev_dev_mutex);
+ netdata_mutex_unlock(&netdev_mutex);
buffer_json_array_close(wb); // data
buffer_json_member_add_object(wb, "columns");
@@ -910,49 +768,26 @@ int netdev_function_net_interfaces(BUFFER *wb, int timeout __maybe_unused, const
buffer_json_member_add_time_t(wb, "expires", now_realtime_sec() + 1);
buffer_json_finalize(wb);
- int response = HTTP_RESP_OK;
- if(is_cancelled_cb && is_cancelled_cb(is_cancelled_cb_data)) {
- buffer_flush(wb);
- response = HTTP_RESP_CLIENT_CLOSED_REQUEST;
- }
-
- if(result_cb)
- result_cb(wb, response, result_cb_data);
-
- return response;
+ return HTTP_RESP_OK;
}
// netdev data collection
static void netdev_cleanup() {
- if(likely(netdev_found == netdev_added)) return;
-
- netdev_added = 0;
- struct netdev *d = netdev_root, *last = NULL;
+ struct netdev *d = netdev_root;
while(d) {
if(unlikely(!d->updated)) {
- // collector_info("Removing network device '%s', linked after '%s'", d->name, last?last->name:"ROOT");
-
- if(netdev_last_used == d)
- netdev_last_used = last;
-
- struct netdev *t = d;
+ struct netdev *next = d->next; // keep the next, to continue;
- if(d == netdev_root || !last)
- netdev_root = d = d->next;
-
- else
- last->next = d = d->next;
+ DOUBLE_LINKED_LIST_REMOVE_ITEM_UNSAFE(netdev_root, d, prev, next);
- t->next = NULL;
- netdev_free(t);
- }
- else {
- netdev_added++;
- last = d;
- d->updated = 0;
- d = d->next;
+ netdev_free(d);
+ d = next;
+ continue;
}
+
+ d->updated = false;
+ d = d->next;
}
}
@@ -962,19 +797,9 @@ static struct netdev *get_netdev(const char *name) {
uint32_t hash = simple_hash(name);
// search it, from the last position to the end
- for(d = netdev_last_used ; d ; d = d->next) {
- if(unlikely(hash == d->hash && !strcmp(name, d->name))) {
- netdev_last_used = d->next;
- return d;
- }
- }
-
- // search it from the beginning to the last position we used
- for(d = netdev_root ; d != netdev_last_used ; d = d->next) {
- if(unlikely(hash == d->hash && !strcmp(name, d->name))) {
- netdev_last_used = d->next;
+ for(d = netdev_root ; d ; d = d->next) {
+ if(unlikely(hash == d->hash && !strcmp(name, d->name)))
return d;
- }
}
// create a new one
@@ -984,6 +809,7 @@ static struct netdev *get_netdev(const char *name) {
d->len = strlen(d->name);
d->chart_labels = rrdlabels_create();
d->function_ready = false;
+ d->double_linked = false;
d->chart_type_net_bytes = strdupz("net");
d->chart_type_net_compressed = strdupz("net_compressed");
@@ -1027,23 +853,26 @@ static struct netdev *get_netdev(const char *name) {
d->chart_family = strdupz(d->name);
d->priority = NETDATA_CHART_PRIO_FIRST_NET_IFACE;
- netdev_rename_lock(d);
-
- netdev_added++;
-
- // link it to the end
- if(netdev_root) {
- struct netdev *e;
- for(e = netdev_root; e->next ; e = e->next) ;
- e->next = d;
- }
- else
- netdev_root = d;
+ DOUBLE_LINKED_LIST_APPEND_ITEM_UNSAFE(netdev_root, d, prev, next);
return d;
}
-#define NETDEV_VIRTUAL_COLLECT_DELAY 15 // 1 full run of the cgroups discovery thread (10 secs by default)
+static bool is_iface_double_linked(struct netdev *d) {
+ char filename[FILENAME_MAX + 1];
+ unsigned long long iflink = 0;
+ unsigned long long ifindex = 0;
+
+ snprintfz(filename, FILENAME_MAX, "%s/sys/class/net/%s/iflink", netdata_configured_host_prefix, d->name);
+ if (read_single_number_file(filename, &iflink))
+ return false;
+
+ snprintfz(filename, FILENAME_MAX, "%s/sys/class/net/%s/ifindex", netdata_configured_host_prefix, d->name);
+ if (read_single_number_file(filename, &ifindex))
+ return false;
+
+ return iflink != ifindex;
+}
int do_proc_net_dev(int update_every, usec_t dt) {
(void)dt;
@@ -1102,6 +931,8 @@ int do_proc_net_dev(int update_every, usec_t dt) {
disabled_list = simple_pattern_create(
config_get(CONFIG_SECTION_PLUGIN_PROC_NETDEV, "disable by default interfaces matching",
"lo fireqos* *-ifb fwpr* fwbr* fwln*"), NULL, SIMPLE_PATTERN_EXACT, true);
+
+ netdev_renames_init();
}
if(unlikely(!ff)) {
@@ -1112,12 +943,6 @@ int do_proc_net_dev(int update_every, usec_t dt) {
ff = procfile_readall(ff);
if(unlikely(!ff)) return 0; // we return 0, so that we will retry to open it next time
- // rename all the devices, if we have pending renames
- if(unlikely(netdev_pending_renames))
- netdev_rename_all_lock();
-
- netdev_found = 0;
-
kernel_uint_t system_rbytes = 0;
kernel_uint_t system_tbytes = 0;
@@ -1133,14 +958,13 @@ int do_proc_net_dev(int update_every, usec_t dt) {
if(name[len - 1] == ':') name[len - 1] = '\0';
struct netdev *d = get_netdev(name);
- d->updated = 1;
- netdev_found++;
+ d->updated = true;
if(unlikely(!d->configured)) {
- // this is the first time we see this interface
+ // the first time we see this interface
// remember we configured it
- d->configured = 1;
+ d->configured = true;
d->discover_time = now;
d->enabled = enable_new_interfaces;
@@ -1151,12 +975,12 @@ int do_proc_net_dev(int update_every, usec_t dt) {
char buf[FILENAME_MAX + 1];
snprintfz(buf, FILENAME_MAX, path_to_sys_devices_virtual_net, d->name);
- d->virtual = likely(access(buf, R_OK) == 0) ? 1 : 0;
+ d->virtual = likely(access(buf, R_OK) == 0) ? true : false;
// At least on Proxmox inside LXC: eth0 is virtual.
// Virtual interfaces are not taken into account in system.net calculations
if (inside_lxc_container && d->virtual && strncmp(d->name, "eth", 3) == 0)
- d->virtual = 0;
+ d->virtual = false;
if (d->virtual)
rrdlabels_add(d->chart_labels, "interface_type", "virtual", RRDLABEL_SRC_AUTO);
@@ -1193,6 +1017,8 @@ int do_proc_net_dev(int update_every, usec_t dt) {
if(d->enabled == CONFIG_BOOLEAN_NO)
continue;
+ d->double_linked = is_iface_double_linked(d);
+
d->do_bandwidth = do_bandwidth;
d->do_packets = do_packets;
d->do_errors = do_errors;
@@ -1235,13 +1061,15 @@ int do_proc_net_dev(int update_every, usec_t dt) {
if(unlikely(!d->enabled))
continue;
+ if(!d->cgroup_netdev_link)
+ netdev_rename_this_device(d);
+
// See https://github.com/netdata/netdata/issues/15206
// This is necessary to prevent the creation of charts for virtual interfaces that will later be
// recreated as container interfaces (create container) or
// rediscovered and recreated only to be deleted almost immediately (stop/remove container)
- if (d->virtual && (now - d->discover_time < NETDEV_VIRTUAL_COLLECT_DELAY)) {
+ if (d->double_linked && d->virtual && (now - d->discover_time < double_linked_device_collect_delay_secs))
continue;
- }
if(likely(d->do_bandwidth != CONFIG_BOOLEAN_NO || !d->virtual)) {
d->rbytes = str2kernel_uint_t(procfile_lineword(ff, l, 1));
@@ -1313,7 +1141,7 @@ int do_proc_net_dev(int update_every, usec_t dt) {
now_monotonic_sec() - d->duplex_file_lost_time > READ_RETRY_PERIOD)) {
char buffer[STATE_LENGTH_MAX + 1];
- if (read_file(d->filename_duplex, buffer, STATE_LENGTH_MAX)) {
+ if (read_txt_file(d->filename_duplex, buffer, sizeof(buffer))) {
if (d->duplex_file_exists)
collector_error("Cannot refresh interface %s duplex state by reading '%s'.", d->name, d->filename_duplex);
d->duplex_file_exists = 0;
@@ -1337,7 +1165,7 @@ int do_proc_net_dev(int update_every, usec_t dt) {
if(d->do_operstate != CONFIG_BOOLEAN_NO && d->filename_operstate) {
char buffer[STATE_LENGTH_MAX + 1], *trimmed_buffer;
- if (read_file(d->filename_operstate, buffer, STATE_LENGTH_MAX)) {
+ if (read_txt_file(d->filename_operstate, buffer, sizeof(buffer))) {
collector_error(
"Cannot refresh %s operstate by reading '%s'. Will not update its status anymore.",
d->name, d->filename_operstate);
@@ -1414,20 +1242,20 @@ int do_proc_net_dev(int update_every, usec_t dt) {
d->flipped ? d->rd_tbytes->collector.last_stored_value : -d->rd_rbytes->collector.last_stored_value,
d->flipped ? -d->rd_rbytes->collector.last_stored_value : d->rd_tbytes->collector.last_stored_value);
- // update the interface speed
- if(d->filename_speed) {
- if(unlikely(!d->chart_var_speed)) {
- d->chart_var_speed =
- rrdsetvar_custom_chart_variable_add_and_acquire(d->st_bandwidth, "nic_speed_max");
- if(!d->chart_var_speed) {
- collector_error(
- "Cannot create interface %s chart variable 'nic_speed_max'. Will not update its speed anymore.",
- d->name);
- freez(d->filename_speed);
- d->filename_speed = NULL;
- }
+ if(unlikely(!d->chart_var_speed)) {
+ d->chart_var_speed = rrdvar_chart_variable_add_and_acquire(d->st_bandwidth, "nic_speed_max");
+ if(!d->chart_var_speed) {
+ collector_error(
+ "Cannot create interface %s chart variable 'nic_speed_max'. Will not update its speed anymore.",
+ d->name);
+ }
+ else {
+ rrdvar_chart_variable_set(d->st_bandwidth, d->chart_var_speed, NAN);
}
+ }
+ // update the interface speed
+ if(d->filename_speed) {
if (d->filename_speed && d->chart_var_speed) {
int ret = 0;
@@ -1473,7 +1301,7 @@ int do_proc_net_dev(int update_every, usec_t dt) {
rrdset_done(d->st_speed);
}
- rrdsetvar_custom_chart_variable_set(
+ rrdvar_chart_variable_set(
d->st_bandwidth, d->chart_var_speed, (NETDATA_DOUBLE)d->speed * KILOBITS_IN_A_MEGABIT);
if (d->speed) {
@@ -1924,32 +1752,36 @@ void *netdev_main(void *ptr)
worker_register("NETDEV");
worker_register_job_name(0, "netdev");
- netdata_thread_cleanup_push(netdev_main_cleanup, ptr);
+ if (getenv("KUBERNETES_SERVICE_HOST") != NULL && getenv("KUBERNETES_SERVICE_PORT") != NULL)
+ double_linked_device_collect_delay_secs = 300;
- rrd_collector_started();
- rrd_function_add(localhost, NULL, "network-interfaces", 10, RRDFUNCTIONS_NETDEV_HELP, true, netdev_function_net_interfaces, NULL);
+ rrd_function_add_inline(localhost, NULL, "network-interfaces", 10,
+ RRDFUNCTIONS_PRIORITY_DEFAULT, RRDFUNCTIONS_NETDEV_HELP,
+ "top", HTTP_ACCESS_ANONYMOUS_DATA,
+ netdev_function_net_interfaces);
- usec_t step = localhost->rrd_update_every * USEC_PER_SEC;
- heartbeat_t hb;
- heartbeat_init(&hb);
+ netdata_thread_cleanup_push(netdev_main_cleanup, ptr) {
+ usec_t step = localhost->rrd_update_every * USEC_PER_SEC;
+ heartbeat_t hb;
+ heartbeat_init(&hb);
- while (service_running(SERVICE_COLLECTORS)) {
- worker_is_idle();
- usec_t hb_dt = heartbeat_next(&hb, step);
+ while (service_running(SERVICE_COLLECTORS)) {
+ worker_is_idle();
+ usec_t hb_dt = heartbeat_next(&hb, step);
- if (unlikely(!service_running(SERVICE_COLLECTORS)))
- break;
+ if (unlikely(!service_running(SERVICE_COLLECTORS)))
+ break;
- cgroup_netdev_reset_all();
+ cgroup_netdev_reset_all();
- worker_is_busy(0);
+ worker_is_busy(0);
- netdata_mutex_lock(&netdev_dev_mutex);
- if(do_proc_net_dev(localhost->rrd_update_every, hb_dt))
- break;
- netdata_mutex_unlock(&netdev_dev_mutex);
+ netdata_mutex_lock(&netdev_mutex);
+ if (do_proc_net_dev(localhost->rrd_update_every, hb_dt))
+ break;
+ netdata_mutex_unlock(&netdev_mutex);
+ }
}
-
netdata_thread_cleanup_pop(1);
return NULL;
diff --git a/collectors/proc.plugin/proc_net_ip_vs_stats.c b/src/collectors/proc.plugin/proc_net_ip_vs_stats.c
index 2b9c9332e..2b9c9332e 100644
--- a/collectors/proc.plugin/proc_net_ip_vs_stats.c
+++ b/src/collectors/proc.plugin/proc_net_ip_vs_stats.c
diff --git a/collectors/proc.plugin/proc_net_netstat.c b/src/collectors/proc.plugin/proc_net_netstat.c
index 170daad5d..4a999803f 100644
--- a/collectors/proc.plugin/proc_net_netstat.c
+++ b/src/collectors/proc.plugin/proc_net_netstat.c
@@ -1647,7 +1647,7 @@ int do_proc_net_netstat(int update_every, usec_t dt) {
arl_expect(arl_udplite, "InCsumErrors", &snmp_root.udplite_InCsumErrors);
arl_expect(arl_udplite, "IgnoredMulti", &snmp_root.udplite_IgnoredMulti);
- tcp_max_connections_var = rrdvar_custom_host_variable_add_and_acquire(localhost, "tcp_max_connections");
+ tcp_max_connections_var = rrdvar_host_variable_add_and_acquire(localhost, "tcp_max_connections");
}
size_t lines, l, words;
@@ -2713,7 +2713,7 @@ int do_proc_net_netstat(int update_every, usec_t dt) {
// snmp Tcp charts
// this is smart enough to update it, only when it is changed
- rrdvar_custom_host_variable_set(localhost, tcp_max_connections_var, snmp_root.tcp_MaxConn);
+ rrdvar_host_variable_set(localhost, tcp_max_connections_var, snmp_root.tcp_MaxConn);
// see http://net-snmp.sourceforge.net/docs/mibs/tcp.html
if(do_tcp_sockets == CONFIG_BOOLEAN_YES || (do_tcp_sockets == CONFIG_BOOLEAN_AUTO &&
diff --git a/collectors/proc.plugin/proc_net_rpc_nfs.c b/src/collectors/proc.plugin/proc_net_rpc_nfs.c
index d6547636e..d6547636e 100644
--- a/collectors/proc.plugin/proc_net_rpc_nfs.c
+++ b/src/collectors/proc.plugin/proc_net_rpc_nfs.c
diff --git a/collectors/proc.plugin/proc_net_rpc_nfsd.c b/src/collectors/proc.plugin/proc_net_rpc_nfsd.c
index 1d9127a03..1d9127a03 100644
--- a/collectors/proc.plugin/proc_net_rpc_nfsd.c
+++ b/src/collectors/proc.plugin/proc_net_rpc_nfsd.c
diff --git a/collectors/proc.plugin/proc_net_sctp_snmp.c b/src/collectors/proc.plugin/proc_net_sctp_snmp.c
index e67143e69..e67143e69 100644
--- a/collectors/proc.plugin/proc_net_sctp_snmp.c
+++ b/src/collectors/proc.plugin/proc_net_sctp_snmp.c
diff --git a/collectors/proc.plugin/proc_net_sockstat.c b/src/collectors/proc.plugin/proc_net_sockstat.c
index b0feab5fa..4be67d61e 100644
--- a/collectors/proc.plugin/proc_net_sockstat.c
+++ b/src/collectors/proc.plugin/proc_net_sockstat.c
@@ -32,9 +32,9 @@ static int read_tcp_mem(void) {
*tcp_mem_high_threshold = NULL;
if(unlikely(!tcp_mem_low_threshold)) {
- tcp_mem_low_threshold = rrdvar_custom_host_variable_add_and_acquire(localhost, "tcp_mem_low");
- tcp_mem_pressure_threshold = rrdvar_custom_host_variable_add_and_acquire(localhost, "tcp_mem_pressure");
- tcp_mem_high_threshold = rrdvar_custom_host_variable_add_and_acquire(localhost, "tcp_mem_high");
+ tcp_mem_low_threshold = rrdvar_host_variable_add_and_acquire(localhost, "tcp_mem_low");
+ tcp_mem_pressure_threshold = rrdvar_host_variable_add_and_acquire(localhost, "tcp_mem_pressure");
+ tcp_mem_high_threshold = rrdvar_host_variable_add_and_acquire(localhost, "tcp_mem_high");
}
if(unlikely(!filename)) {
@@ -44,7 +44,7 @@ static int read_tcp_mem(void) {
}
char buffer[200 + 1], *start, *end;
- if(read_file(filename, buffer, 200) != 0) return 1;
+ if(read_txt_file(filename, buffer, sizeof(buffer)) != 0) return 1;
buffer[200] = '\0';
unsigned long long low = 0, pressure = 0, high = 0;
@@ -60,9 +60,9 @@ static int read_tcp_mem(void) {
// fprintf(stderr, "TCP MEM low = %llu, pressure = %llu, high = %llu\n", low, pressure, high);
- rrdvar_custom_host_variable_set(localhost, tcp_mem_low_threshold, low * sysconf(_SC_PAGESIZE) / 1024.0);
- rrdvar_custom_host_variable_set(localhost, tcp_mem_pressure_threshold, pressure * sysconf(_SC_PAGESIZE) / 1024.0);
- rrdvar_custom_host_variable_set(localhost, tcp_mem_high_threshold, high * sysconf(_SC_PAGESIZE) / 1024.0);
+ rrdvar_host_variable_set(localhost, tcp_mem_low_threshold, low * sysconf(_SC_PAGESIZE) / 1024.0);
+ rrdvar_host_variable_set(localhost, tcp_mem_pressure_threshold, pressure * sysconf(_SC_PAGESIZE) / 1024.0);
+ rrdvar_host_variable_set(localhost, tcp_mem_high_threshold, high * sysconf(_SC_PAGESIZE) / 1024.0);
return 0;
}
@@ -81,9 +81,9 @@ static kernel_uint_t read_tcp_max_orphans(void) {
if(read_single_number_file(filename, &tcp_max_orphans) == 0) {
if(unlikely(!tcp_max_orphans_var))
- tcp_max_orphans_var = rrdvar_custom_host_variable_add_and_acquire(localhost, "tcp_max_orphans");
+ tcp_max_orphans_var = rrdvar_host_variable_add_and_acquire(localhost, "tcp_max_orphans");
- rrdvar_custom_host_variable_set(localhost, tcp_max_orphans_var, tcp_max_orphans);
+ rrdvar_host_variable_set(localhost, tcp_max_orphans_var, tcp_max_orphans);
return tcp_max_orphans;
}
diff --git a/collectors/proc.plugin/proc_net_sockstat6.c b/src/collectors/proc.plugin/proc_net_sockstat6.c
index 16e0248af..16e0248af 100644
--- a/collectors/proc.plugin/proc_net_sockstat6.c
+++ b/src/collectors/proc.plugin/proc_net_sockstat6.c
diff --git a/collectors/proc.plugin/proc_net_softnet_stat.c b/src/collectors/proc.plugin/proc_net_softnet_stat.c
index 2f01b8859..a225a9f0d 100644
--- a/collectors/proc.plugin/proc_net_softnet_stat.c
+++ b/src/collectors/proc.plugin/proc_net_softnet_stat.c
@@ -24,7 +24,10 @@ int do_proc_net_softnet_stat(int update_every, usec_t dt) {
static size_t allocated_lines = 0, allocated_columns = 0;
static uint32_t *data = NULL;
- if(unlikely(do_per_core == -1)) do_per_core = config_get_boolean("plugin:proc:/proc/net/softnet_stat", "softnet_stat per core", 1);
+ if (unlikely(do_per_core == -1)) {
+ do_per_core =
+ config_get_boolean("plugin:proc:/proc/net/softnet_stat", "softnet_stat per core", CONFIG_BOOLEAN_NO);
+ }
if(unlikely(!ff)) {
char filename[FILENAME_MAX + 1];
diff --git a/collectors/proc.plugin/proc_net_stat_conntrack.c b/src/collectors/proc.plugin/proc_net_stat_conntrack.c
index e8fbdbb66..6951cba79 100644
--- a/collectors/proc.plugin/proc_net_stat_conntrack.c
+++ b/src/collectors/proc.plugin/proc_net_stat_conntrack.c
@@ -50,7 +50,7 @@ int do_proc_net_stat_conntrack(int update_every, usec_t dt) {
if(!do_sockets && !read_full)
return 1;
- rrdvar_max = rrdvar_custom_host_variable_add_and_acquire(localhost, "netfilter_conntrack_max");
+ rrdvar_max = rrdvar_host_variable_add_and_acquire(localhost, "netfilter_conntrack_max");
}
if(likely(read_full)) {
@@ -125,7 +125,7 @@ int do_proc_net_stat_conntrack(int update_every, usec_t dt) {
unsigned long long max;
if(likely(!read_single_number_file(nf_conntrack_max_filename, &max)))
- rrdvar_custom_host_variable_set(localhost, rrdvar_max, max);
+ rrdvar_host_variable_set(localhost, rrdvar_max, max);
}
// --------------------------------------------------------------------
diff --git a/collectors/proc.plugin/proc_net_stat_synproxy.c b/src/collectors/proc.plugin/proc_net_stat_synproxy.c
index e23a0ab7b..e23a0ab7b 100644
--- a/collectors/proc.plugin/proc_net_stat_synproxy.c
+++ b/src/collectors/proc.plugin/proc_net_stat_synproxy.c
diff --git a/collectors/proc.plugin/proc_net_wireless.c b/src/collectors/proc.plugin/proc_net_wireless.c
index c7efa3335..c7efa3335 100644
--- a/collectors/proc.plugin/proc_net_wireless.c
+++ b/src/collectors/proc.plugin/proc_net_wireless.c
diff --git a/collectors/proc.plugin/proc_pagetypeinfo.c b/src/collectors/proc.plugin/proc_pagetypeinfo.c
index fc5496c63..fc5496c63 100644
--- a/collectors/proc.plugin/proc_pagetypeinfo.c
+++ b/src/collectors/proc.plugin/proc_pagetypeinfo.c
diff --git a/collectors/proc.plugin/proc_pressure.c b/src/collectors/proc.plugin/proc_pressure.c
index 4037e60ac..4037e60ac 100644
--- a/collectors/proc.plugin/proc_pressure.c
+++ b/src/collectors/proc.plugin/proc_pressure.c
diff --git a/collectors/proc.plugin/proc_pressure.h b/src/collectors/proc.plugin/proc_pressure.h
index 2e5cab2cc..2e5cab2cc 100644
--- a/collectors/proc.plugin/proc_pressure.h
+++ b/src/collectors/proc.plugin/proc_pressure.h
diff --git a/collectors/proc.plugin/proc_self_mountinfo.c b/src/collectors/proc.plugin/proc_self_mountinfo.c
index 194791603..194791603 100644
--- a/collectors/proc.plugin/proc_self_mountinfo.c
+++ b/src/collectors/proc.plugin/proc_self_mountinfo.c
diff --git a/collectors/proc.plugin/proc_self_mountinfo.h b/src/collectors/proc.plugin/proc_self_mountinfo.h
index 4bd24d2d2..4bd24d2d2 100644
--- a/collectors/proc.plugin/proc_self_mountinfo.h
+++ b/src/collectors/proc.plugin/proc_self_mountinfo.h
diff --git a/collectors/proc.plugin/proc_softirqs.c b/src/collectors/proc.plugin/proc_softirqs.c
index 5f0502f66..7968a2287 100644
--- a/collectors/proc.plugin/proc_softirqs.c
+++ b/src/collectors/proc.plugin/proc_softirqs.c
@@ -59,7 +59,7 @@ int do_proc_softirqs(int update_every, usec_t dt) {
struct interrupt *irrs = NULL;
if(unlikely(do_per_core == CONFIG_BOOLEAN_INVALID))
- do_per_core = config_get_boolean_ondemand("plugin:proc:/proc/softirqs", "interrupts per core", CONFIG_BOOLEAN_AUTO);
+ do_per_core = config_get_boolean_ondemand("plugin:proc:/proc/softirqs", "interrupts per core", CONFIG_BOOLEAN_NO);
if(unlikely(!ff)) {
char filename[FILENAME_MAX + 1];
diff --git a/collectors/proc.plugin/proc_spl_kstat_zfs.c b/src/collectors/proc.plugin/proc_spl_kstat_zfs.c
index 27178b60f..e6b12c31f 100644
--- a/collectors/proc.plugin/proc_spl_kstat_zfs.c
+++ b/src/collectors/proc.plugin/proc_spl_kstat_zfs.c
@@ -6,7 +6,7 @@
#define ZFS_PROC_ARCSTATS "/proc/spl/kstat/zfs/arcstats"
#define ZFS_PROC_POOLS "/proc/spl/kstat/zfs"
-#define STATE_SIZE 9
+#define STATE_SIZE 20
#define MAX_CHART_ID 256
extern struct arcstats arcstats;
@@ -272,7 +272,7 @@ int update_zfs_pool_state_chart(const DICTIONARY_ITEM *item, void *pool_p, void
"zfspool",
chart_id,
NULL,
- name,
+ "state",
"zfspool.state",
"ZFS pool state",
"boolean",
@@ -357,7 +357,7 @@ int do_proc_spl_kstat_zfs_pool_state(int update_every, usec_t dt)
if (unlikely(!pool)) {
struct zfs_pool new_zfs_pool = {};
pool = dictionary_set(zfs_pools, de->d_name, &new_zfs_pool, sizeof(struct zfs_pool));
- };
+ }
pool->updated = 1;
@@ -378,7 +378,7 @@ int do_proc_spl_kstat_zfs_pool_state(int update_every, usec_t dt)
snprintfz(filename, FILENAME_MAX, "%s/%s/state", dirname, de->d_name);
char state[STATE_SIZE + 1];
- int ret = read_file(filename, state, STATE_SIZE);
+ int ret = read_txt_file(filename, state, sizeof(state));
if (!ret) {
state_file_found = 1;
diff --git a/collectors/proc.plugin/proc_stat.c b/src/collectors/proc.plugin/proc_stat.c
index 84160f22f..481cb906a 100644
--- a/collectors/proc.plugin/proc_stat.c
+++ b/src/collectors/proc.plugin/proc_stat.c
@@ -46,6 +46,7 @@ struct cpu_chart {
RRDDIM *rd_guest;
RRDDIM *rd_guest_nice;
+ bool per_core_files_found;
struct per_core_single_number_file files[PER_CORE_FILES];
struct per_core_time_in_state_file time_in_state_files;
@@ -67,7 +68,7 @@ static int read_per_core_files(struct cpu_chart *all_cpu_charts, size_t len, siz
continue;
if(unlikely(f->fd == -1)) {
- f->fd = open(f->filename, O_RDONLY);
+ f->fd = open(f->filename, O_RDONLY | O_CLOEXEC);
if (unlikely(f->fd == -1)) {
collector_error("Cannot open file '%s'", f->filename);
continue;
@@ -411,7 +412,7 @@ static int read_cpuidle_states(char *cpuidle_name_filename , char *cpuidle_time_
char name_buf[50 + 1];
snprintfz(filename, FILENAME_MAX, cpuidle_name_filename, core, state);
- int fd = open(filename, O_RDONLY, 0666);
+ int fd = open(filename, O_RDONLY | O_CLOEXEC, 0666);
if(unlikely(fd == -1)) {
collector_error("Cannot open file '%s'", filename);
cc->rescan_cpu_states = 1;
@@ -443,7 +444,7 @@ static int read_cpuidle_states(char *cpuidle_name_filename , char *cpuidle_time_
struct cpuidle_state *cs = &cc->cpuidle_state[state];
if(unlikely(cs->time_fd == -1)) {
- cs->time_fd = open(cs->time_filename, O_RDONLY);
+ cs->time_fd = open(cs->time_filename, O_RDONLY | O_CLOEXEC);
if (unlikely(cs->time_fd == -1)) {
collector_error("Cannot open file '%s'", cs->time_filename);
cc->rescan_cpu_states = 1;
@@ -487,7 +488,7 @@ int do_proc_stat(int update_every, usec_t dt) {
if(unlikely(do_cpu == -1)) {
do_cpu = config_get_boolean("plugin:proc:/proc/stat", "cpu utilization", CONFIG_BOOLEAN_YES);
- do_cpu_cores = config_get_boolean("plugin:proc:/proc/stat", "per cpu core utilization", CONFIG_BOOLEAN_YES);
+ do_cpu_cores = config_get_boolean("plugin:proc:/proc/stat", "per cpu core utilization", CONFIG_BOOLEAN_NO);
do_interrupts = config_get_boolean("plugin:proc:/proc/stat", "cpu interrupts", CONFIG_BOOLEAN_YES);
do_context = config_get_boolean("plugin:proc:/proc/stat", "context switches", CONFIG_BOOLEAN_YES);
do_forks = config_get_boolean("plugin:proc:/proc/stat", "processes started", CONFIG_BOOLEAN_YES);
@@ -508,7 +509,7 @@ int do_proc_stat(int update_every, usec_t dt) {
do_core_throttle_count = CONFIG_BOOLEAN_AUTO;
do_package_throttle_count = CONFIG_BOOLEAN_NO;
do_cpu_freq = CONFIG_BOOLEAN_YES;
- do_cpuidle = CONFIG_BOOLEAN_YES;
+ do_cpuidle = CONFIG_BOOLEAN_NO;
}
if(unlikely(get_system_cpus() > 24)) {
// the system has too many processors
@@ -589,14 +590,71 @@ int do_proc_stat(int update_every, usec_t dt) {
continue;
}
- size_t core = (row_key[3] == '\0') ? 0 : str2ul(&row_key[3]) + 1;
- if(likely(core > 0)) cores_found = core;
+ size_t core = (row_key[3] == '\0') ? 0 : str2ul(&row_key[3]) + 1;
+ if (likely(core > 0))
+ cores_found = core;
+
+ bool do_any_core_metric = do_cpu_cores || do_core_throttle_count || do_cpu_freq || do_cpuidle;
+
+ if (likely((core == 0 && do_cpu) || (core > 0 && do_any_core_metric))) {
+ if (unlikely(core >= all_cpu_charts_size)) {
+ size_t old_cpu_charts_size = all_cpu_charts_size;
+ all_cpu_charts_size = core + 1;
+ all_cpu_charts = reallocz(all_cpu_charts, sizeof(struct cpu_chart) * all_cpu_charts_size);
+ memset(&all_cpu_charts[old_cpu_charts_size], 0, sizeof(struct cpu_chart) * (all_cpu_charts_size - old_cpu_charts_size));
+ }
+
+ struct cpu_chart *cpu_chart = &all_cpu_charts[core];
+
+ if (unlikely(!cpu_chart->id))
+ cpu_chart->id = strdupz(row_key);
+
+ if (core > 0 && !cpu_chart->per_core_files_found) {
+ cpu_chart->per_core_files_found = true;
+
+ char filename[FILENAME_MAX + 1];
+ struct stat stbuf;
+
+ if (do_core_throttle_count != CONFIG_BOOLEAN_NO) {
+ snprintfz(filename, FILENAME_MAX, core_throttle_count_filename, cpu_chart->id);
+ if (stat(filename, &stbuf) == 0) {
+ cpu_chart->files[CORE_THROTTLE_COUNT_INDEX].filename = strdupz(filename);
+ cpu_chart->files[CORE_THROTTLE_COUNT_INDEX].fd = -1;
+ do_core_throttle_count = CONFIG_BOOLEAN_YES;
+ }
+ }
+
+ if (do_package_throttle_count != CONFIG_BOOLEAN_NO) {
+ snprintfz(filename, FILENAME_MAX, package_throttle_count_filename, cpu_chart->id);
+ if (stat(filename, &stbuf) == 0) {
+ cpu_chart->files[PACKAGE_THROTTLE_COUNT_INDEX].filename = strdupz(filename);
+ cpu_chart->files[PACKAGE_THROTTLE_COUNT_INDEX].fd = -1;
+ do_package_throttle_count = CONFIG_BOOLEAN_YES;
+ }
+ }
+
+ if (do_cpu_freq != CONFIG_BOOLEAN_NO) {
+ snprintfz(filename, FILENAME_MAX, scaling_cur_freq_filename, cpu_chart->id);
+ if (stat(filename, &stbuf) == 0) {
+ cpu_chart->files[CPU_FREQ_INDEX].filename = strdupz(filename);
+ cpu_chart->files[CPU_FREQ_INDEX].fd = -1;
+ do_cpu_freq = CONFIG_BOOLEAN_YES;
+ }
+
+ snprintfz(filename, FILENAME_MAX, time_in_state_filename, cpu_chart->id);
+ if (stat(filename, &stbuf) == 0) {
+ cpu_chart->time_in_state_files.filename = strdupz(filename);
+ cpu_chart->time_in_state_files.ff = NULL;
+ do_cpu_freq = CONFIG_BOOLEAN_YES;
+ accurate_freq_avail = 1;
+ }
+ }
+ }
+ }
if(likely((core == 0 && do_cpu) || (core > 0 && do_cpu_cores))) {
- char *id;
unsigned long long user = 0, nice = 0, system = 0, idle = 0, iowait = 0, irq = 0, softirq = 0, steal = 0, guest = 0, guest_nice = 0;
- id = row_key;
user = str2ull(procfile_lineword(ff, l, 1), NULL);
nice = str2ull(procfile_lineword(ff, l, 2), NULL);
system = str2ull(procfile_lineword(ff, l, 3), NULL);
@@ -615,17 +673,11 @@ int do_proc_stat(int update_every, usec_t dt) {
char *title, *type, *context, *family;
long priority;
- if(unlikely(core >= all_cpu_charts_size)) {
- size_t old_cpu_charts_size = all_cpu_charts_size;
- all_cpu_charts_size = core + 1;
- all_cpu_charts = reallocz(all_cpu_charts, sizeof(struct cpu_chart) * all_cpu_charts_size);
- memset(&all_cpu_charts[old_cpu_charts_size], 0, sizeof(struct cpu_chart) * (all_cpu_charts_size - old_cpu_charts_size));
- }
struct cpu_chart *cpu_chart = &all_cpu_charts[core];
- if(unlikely(!cpu_chart->st)) {
- cpu_chart->id = strdupz(id);
+ char *id = row_key;
+ if(unlikely(!cpu_chart->st)) {
if(unlikely(core == 0)) {
title = "Total CPU utilization";
type = "system";
@@ -639,47 +691,6 @@ int do_proc_stat(int update_every, usec_t dt) {
context = "cpu.cpu";
family = "utilization";
priority = NETDATA_CHART_PRIO_CPU_PER_CORE;
-
- char filename[FILENAME_MAX + 1];
- struct stat stbuf;
-
- if(do_core_throttle_count != CONFIG_BOOLEAN_NO) {
- snprintfz(filename, FILENAME_MAX, core_throttle_count_filename, id);
- if (stat(filename, &stbuf) == 0) {
- cpu_chart->files[CORE_THROTTLE_COUNT_INDEX].filename = strdupz(filename);
- cpu_chart->files[CORE_THROTTLE_COUNT_INDEX].fd = -1;
- do_core_throttle_count = CONFIG_BOOLEAN_YES;
- }
- }
-
- if(do_package_throttle_count != CONFIG_BOOLEAN_NO) {
- snprintfz(filename, FILENAME_MAX, package_throttle_count_filename, id);
- if (stat(filename, &stbuf) == 0) {
- cpu_chart->files[PACKAGE_THROTTLE_COUNT_INDEX].filename = strdupz(filename);
- cpu_chart->files[PACKAGE_THROTTLE_COUNT_INDEX].fd = -1;
- do_package_throttle_count = CONFIG_BOOLEAN_YES;
- }
- }
-
- if(do_cpu_freq != CONFIG_BOOLEAN_NO) {
-
- snprintfz(filename, FILENAME_MAX, scaling_cur_freq_filename, id);
-
- if (stat(filename, &stbuf) == 0) {
- cpu_chart->files[CPU_FREQ_INDEX].filename = strdupz(filename);
- cpu_chart->files[CPU_FREQ_INDEX].fd = -1;
- do_cpu_freq = CONFIG_BOOLEAN_YES;
- }
-
- snprintfz(filename, FILENAME_MAX, time_in_state_filename, id);
-
- if (stat(filename, &stbuf) == 0) {
- cpu_chart->time_in_state_files.filename = strdupz(filename);
- cpu_chart->time_in_state_files.ff = NULL;
- do_cpu_freq = CONFIG_BOOLEAN_YES;
- accurate_freq_avail = 1;
- }
- }
}
cpu_chart->st = rrdset_create_localhost(
@@ -719,7 +730,7 @@ int do_proc_stat(int update_every, usec_t dt) {
}
if(unlikely(core == 0 && cpus_var == NULL))
- cpus_var = rrdvar_custom_host_variable_add_and_acquire(localhost, "active_processors");
+ cpus_var = rrdvar_host_variable_add_and_acquire(localhost, "active_processors");
}
rrddim_set_by_pointer(cpu_chart->st, cpu_chart->rd_user, user);
@@ -1064,7 +1075,7 @@ int do_proc_stat(int update_every, usec_t dt) {
}
if(cpus_var)
- rrdvar_custom_host_variable_set(localhost, cpus_var, cores_found);
+ rrdvar_host_variable_set(localhost, cpus_var, cores_found);
return 0;
}
diff --git a/collectors/proc.plugin/proc_sys_fs_file_nr.c b/src/collectors/proc.plugin/proc_sys_fs_file_nr.c
index 570945d01..570945d01 100644
--- a/collectors/proc.plugin/proc_sys_fs_file_nr.c
+++ b/src/collectors/proc.plugin/proc_sys_fs_file_nr.c
diff --git a/collectors/proc.plugin/proc_sys_kernel_random_entropy_avail.c b/src/collectors/proc.plugin/proc_sys_kernel_random_entropy_avail.c
index b32597bc4..b32597bc4 100644
--- a/collectors/proc.plugin/proc_sys_kernel_random_entropy_avail.c
+++ b/src/collectors/proc.plugin/proc_sys_kernel_random_entropy_avail.c
diff --git a/collectors/proc.plugin/proc_uptime.c b/src/collectors/proc.plugin/proc_uptime.c
index ddab7269b..ddab7269b 100644
--- a/collectors/proc.plugin/proc_uptime.c
+++ b/src/collectors/proc.plugin/proc_uptime.c
diff --git a/collectors/proc.plugin/proc_vmstat.c b/src/collectors/proc.plugin/proc_vmstat.c
index b44733b6a..b44733b6a 100644
--- a/collectors/proc.plugin/proc_vmstat.c
+++ b/src/collectors/proc.plugin/proc_vmstat.c
diff --git a/collectors/proc.plugin/sys_block_zram.c b/src/collectors/proc.plugin/sys_block_zram.c
index dac7cac0f..dac7cac0f 100644
--- a/collectors/proc.plugin/sys_block_zram.c
+++ b/src/collectors/proc.plugin/sys_block_zram.c
diff --git a/collectors/proc.plugin/sys_class_drm.c b/src/collectors/proc.plugin/sys_class_drm.c
index 3ed1fb875..ab4d98a72 100644
--- a/collectors/proc.plugin/sys_class_drm.c
+++ b/src/collectors/proc.plugin/sys_class_drm.c
@@ -849,8 +849,6 @@ int do_sys_class_drm(int update_every, usec_t dt) {
(de->d_name[0] == '.' && de->d_name[1] == '.' && de->d_name[2] == '\0'))) continue;
if(de->d_type == DT_LNK && !strncmp(de->d_name, "card", 4) && !strchr(de->d_name, '-')) {
- char filename[FILENAME_MAX + 1];
-
snprintfz(filename, FILENAME_MAX, "%s/%s/%s", drm_dir_name, de->d_name, "device/uevent");
if(check_card_is_amdgpu(filename)) continue;
@@ -885,14 +883,14 @@ int do_sys_class_drm(int update_every, usec_t dt) {
collected_number tmp_val;
- #define set_prop_pathname(prop_filename, prop_pathname, p_ff){ \
+ #define set_prop_pathname(prop_filename, prop_pathname, p_ff) do { \
snprintfz(filename, FILENAME_MAX, "%s/%s", c->pathname, prop_filename); \
if((p_ff && !read_clk_freq_file(p_ff, filename, &tmp_val)) || \
!read_single_number_file(filename, (unsigned long long *) &tmp_val)) \
prop_pathname = strdupz(filename); \
else \
collector_info("Cannot read file '%s'", filename); \
- }
+ } while(0)
/* Initialize GPU and VRAM utilization metrics */
diff --git a/collectors/proc.plugin/sys_class_infiniband.c b/src/collectors/proc.plugin/sys_class_infiniband.c
index d12a34513..ff1652ddf 100644
--- a/collectors/proc.plugin/sys_class_infiniband.c
+++ b/src/collectors/proc.plugin/sys_class_infiniband.c
@@ -184,7 +184,7 @@ static struct ibport {
RRDSET *st_hwpackets;
RRDSET *st_hwerrors;
- const RRDSETVAR_ACQUIRED *stv_speed;
+ const RRDVAR_ACQUIRED *stv_speed;
usec_t speed_last_collected_usec;
@@ -470,7 +470,7 @@ int do_sys_class_infiniband(int update_every, usec_t dt)
snprintfz(buffer, FILENAME_MAX, "%s/%s/%s", ports_dirname, port_dent->d_name, "rate");
char buffer_rate[65];
p->width = 4;
- if (read_file(buffer, buffer_rate, 64)) {
+ if (read_txt_file(buffer, buffer_rate, sizeof(buffer_rate))) {
collector_error("Unable to read '%s'", buffer);
} else {
char *buffer_width = strstr(buffer_rate, "(");
@@ -545,14 +545,14 @@ int do_sys_class_infiniband(int update_every, usec_t dt)
// x4 lanes multiplier as per Documentation/ABI/stable/sysfs-class-infiniband
FOREACH_COUNTER_BYTES(GEN_RRD_DIM_ADD_CUSTOM, port, port->width * 8, 1000, RRD_ALGORITHM_INCREMENTAL)
- port->stv_speed = rrdsetvar_custom_chart_variable_add_and_acquire(port->st_bytes, "link_speed");
+ port->stv_speed = rrdvar_chart_variable_add_and_acquire(port->st_bytes, "link_speed");
}
// Link read values to dimensions
FOREACH_COUNTER_BYTES(GEN_RRD_DIM_SETP, port)
// For link speed set only variable
- rrdsetvar_custom_chart_variable_set(port->st_bytes, port->stv_speed, port->speed);
+ rrdvar_chart_variable_set(port->st_bytes, port->stv_speed, port->speed);
rrdset_done(port->st_bytes);
}
diff --git a/collectors/proc.plugin/sys_class_power_supply.c b/src/collectors/proc.plugin/sys_class_power_supply.c
index 3f793b9c6..494a293bc 100644
--- a/collectors/proc.plugin/sys_class_power_supply.c
+++ b/src/collectors/proc.plugin/sys_class_power_supply.c
@@ -245,7 +245,7 @@ int do_sys_class_power_supply(int update_every, usec_t dt) {
char buffer[30 + 1];
if(unlikely(ps->capacity->fd == -1)) {
- ps->capacity->fd = open(ps->capacity->filename, O_RDONLY, 0666);
+ ps->capacity->fd = open(ps->capacity->filename, O_RDONLY | O_CLOEXEC, 0666);
if(unlikely(ps->capacity->fd == -1)) {
collector_error("Cannot open file '%s'", ps->capacity->filename);
power_supply_free(ps);
@@ -290,7 +290,7 @@ int do_sys_class_power_supply(int update_every, usec_t dt) {
char buffer[30 + 1];
if(unlikely(pd->fd == -1)) {
- pd->fd = open(pd->filename, O_RDONLY, 0666);
+ pd->fd = open(pd->filename, O_RDONLY | O_CLOEXEC, 0666);
if(unlikely(pd->fd == -1)) {
collector_error("Cannot open file '%s'", pd->filename);
read_error = 1;
diff --git a/collectors/proc.plugin/sys_devices_pci_aer.c b/src/collectors/proc.plugin/sys_devices_pci_aer.c
index 563ebf051..563ebf051 100644
--- a/collectors/proc.plugin/sys_devices_pci_aer.c
+++ b/src/collectors/proc.plugin/sys_devices_pci_aer.c
diff --git a/collectors/proc.plugin/sys_devices_system_edac_mc.c b/src/collectors/proc.plugin/sys_devices_system_edac_mc.c
index fdaa22cb7..d3db8c044 100644
--- a/collectors/proc.plugin/sys_devices_system_edac_mc.c
+++ b/src/collectors/proc.plugin/sys_devices_system_edac_mc.c
@@ -150,7 +150,7 @@ static kernel_uint_t read_edac_count(struct edac_count *t) {
static bool read_edac_mc_file(const char *mc, const char *filename, char *out, size_t out_size) {
char f[FILENAME_MAX + 1];
snprintfz(f, FILENAME_MAX, "%s/%s/%s", mc_dirname, mc, filename);
- if(read_file(f, out, out_size) != 0) {
+ if(read_txt_file(f, out, out_size) != 0) {
collector_error("EDAC: cannot read file '%s'", f);
return false;
}
@@ -160,7 +160,7 @@ static bool read_edac_mc_file(const char *mc, const char *filename, char *out, s
static bool read_edac_mc_rank_file(const char *mc, const char *rank, const char *filename, char *out, size_t out_size) {
char f[FILENAME_MAX + 1];
snprintfz(f, FILENAME_MAX, "%s/%s/%s/%s", mc_dirname, mc, rank, filename);
- if(read_file(f, out, out_size) != 0) {
+ if(read_txt_file(f, out, out_size) != 0) {
collector_error("EDAC: cannot read file '%s'", f);
return false;
}
@@ -202,9 +202,9 @@ int do_proc_sys_devices_system_edac_mc(int update_every, usec_t dt __maybe_unuse
, id
, NULL
, "edac"
- , "mem.edac_mc"
+ , "mem.edac_mc_errors"
, "Memory Controller (MC) Error Detection And Correction (EDAC) Errors"
- , "errors/s"
+ , "errors"
, PLUGIN_PROC_NAME
, "/sys/devices/system/edac/mc"
, NETDATA_CHART_PRIO_MEM_HW_ECC_CE
@@ -225,10 +225,10 @@ int do_proc_sys_devices_system_edac_mc(int update_every, usec_t dt __maybe_unuse
if(read_edac_mc_file(m->name, "max_location", buffer, 1024))
rrdlabels_add(m->st->rrdlabels, "max_location", buffer, RRDLABEL_SRC_AUTO);
- m->ce.rd = rrddim_add(m->st, "correctable", NULL, 1, 1, RRD_ALGORITHM_INCREMENTAL);
- m->ue.rd = rrddim_add(m->st, "uncorrectable", NULL, 1, 1, RRD_ALGORITHM_INCREMENTAL);
- m->ce_noinfo.rd = rrddim_add(m->st, "correctable_noinfo", NULL, 1, 1, RRD_ALGORITHM_INCREMENTAL);
- m->ue_noinfo.rd = rrddim_add(m->st, "uncorrectable_noinfo", NULL, 1, 1, RRD_ALGORITHM_INCREMENTAL);
+ m->ce.rd = rrddim_add(m->st, "correctable", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
+ m->ue.rd = rrddim_add(m->st, "uncorrectable", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
+ m->ce_noinfo.rd = rrddim_add(m->st, "correctable_noinfo", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
+ m->ue_noinfo.rd = rrddim_add(m->st, "uncorrectable_noinfo", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
}
rrddim_set_by_pointer(m->st, m->ce.rd, (collected_number)m->ce.count);
@@ -250,9 +250,9 @@ int do_proc_sys_devices_system_edac_mc(int update_every, usec_t dt __maybe_unuse
, id
, NULL
, "edac"
- , "mem.edac_mc_dimm"
+ , "mem.edac_mc_dimm_errors"
, "DIMM Error Detection And Correction (EDAC) Errors"
- , "errors/s"
+ , "errors"
, PLUGIN_PROC_NAME
, "/sys/devices/system/edac/mc"
, NETDATA_CHART_PRIO_MEM_HW_ECC_CE + 1
@@ -283,8 +283,8 @@ int do_proc_sys_devices_system_edac_mc(int update_every, usec_t dt __maybe_unuse
if (read_edac_mc_rank_file(m->name, d->name, "size", buffer, 1024))
rrdlabels_add(d->st->rrdlabels, "size", buffer, RRDLABEL_SRC_AUTO);
- d->ce.rd = rrddim_add(d->st, "correctable", NULL, 1, 1, RRD_ALGORITHM_INCREMENTAL);
- d->ue.rd = rrddim_add(d->st, "uncorrectable", NULL, 1, 1, RRD_ALGORITHM_INCREMENTAL);
+ d->ce.rd = rrddim_add(d->st, "correctable", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
+ d->ue.rd = rrddim_add(d->st, "uncorrectable", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
}
rrddim_set_by_pointer(d->st, d->ce.rd, (collected_number)d->ce.count);
diff --git a/collectors/proc.plugin/sys_devices_system_node.c b/src/collectors/proc.plugin/sys_devices_system_node.c
index d6db94a27..d6db94a27 100644
--- a/collectors/proc.plugin/sys_devices_system_node.c
+++ b/src/collectors/proc.plugin/sys_devices_system_node.c
diff --git a/collectors/proc.plugin/sys_fs_btrfs.c b/src/collectors/proc.plugin/sys_fs_btrfs.c
index f1d6fe720..7023484ca 100644
--- a/collectors/proc.plugin/sys_fs_btrfs.c
+++ b/src/collectors/proc.plugin/sys_fs_btrfs.c
@@ -122,7 +122,7 @@ static BTRFS_NODE *nodes = NULL;
static inline int collect_btrfs_error_stats(BTRFS_DEVICE *device){
char buffer[120 + 1];
- int ret = read_file(device->error_stats_filename, buffer, 120);
+ int ret = read_txt_file(device->error_stats_filename, buffer, sizeof(buffer));
if(unlikely(ret)) {
collector_error("BTRFS: failed to read '%s'", device->error_stats_filename);
device->write_errs = 0;
@@ -151,7 +151,7 @@ static inline int collect_btrfs_error_stats(BTRFS_DEVICE *device){
static inline int collect_btrfs_commits_stats(BTRFS_NODE *node, int update_every){
char buffer[120 + 1];
- int ret = read_file(node->commit_stats_filename, buffer, 120);
+ int ret = read_txt_file(node->commit_stats_filename, buffer, sizeof(buffer));
if(unlikely(ret)) {
collector_error("BTRFS: failed to read '%s'", node->commit_stats_filename);
node->commits_total = 0;
@@ -530,7 +530,7 @@ static inline int find_all_btrfs_pools(const char *path, int update_every) {
char label[FILENAME_MAX + 1] = "";
snprintfz(filename, FILENAME_MAX, "%s/%s/label", path, de->d_name);
- if(read_file(filename, label, FILENAME_MAX) != 0) {
+ if(read_txt_file(filename, label, sizeof(label)) != 0) {
collector_error("BTRFS: failed to read '%s'", filename);
btrfs_free_node(node);
continue;
diff --git a/collectors/proc.plugin/sys_kernel_mm_ksm.c b/src/collectors/proc.plugin/sys_kernel_mm_ksm.c
index 45f1ac330..8f43acc93 100644
--- a/collectors/proc.plugin/sys_kernel_mm_ksm.c
+++ b/src/collectors/proc.plugin/sys_kernel_mm_ksm.c
@@ -13,7 +13,7 @@ typedef struct ksm_name_value {
#define PAGES_SHARING 1
#define PAGES_UNSHARED 2
#define PAGES_VOLATILE 3
-#define PAGES_TO_SCAN 4
+// #define PAGES_TO_SCAN 4
KSM_NAME_VALUE values[] = {
[PAGES_SHARED] = { "/sys/kernel/mm/ksm/pages_shared", 0ULL },
diff --git a/collectors/proc.plugin/zfs_common.c b/src/collectors/proc.plugin/zfs_common.c
index cca0ae0e6..cca0ae0e6 100644
--- a/collectors/proc.plugin/zfs_common.c
+++ b/src/collectors/proc.plugin/zfs_common.c
diff --git a/collectors/proc.plugin/zfs_common.h b/src/collectors/proc.plugin/zfs_common.h
index 9d61de2f3..9d61de2f3 100644
--- a/collectors/proc.plugin/zfs_common.h
+++ b/src/collectors/proc.plugin/zfs_common.h
diff --git a/collectors/profile.plugin/README.md b/src/collectors/profile.plugin/README.md
index 1f200fc31..1f200fc31 100644
--- a/collectors/profile.plugin/README.md
+++ b/src/collectors/profile.plugin/README.md
diff --git a/collectors/profile.plugin/plugin_profile.cc b/src/collectors/profile.plugin/plugin_profile.cc
index 5f7b22d25..f97c192a0 100644
--- a/collectors/profile.plugin/plugin_profile.cc
+++ b/src/collectors/profile.plugin/plugin_profile.cc
@@ -14,8 +14,6 @@ extern "C" {
#include <thread>
#include <vector>
-#define PLUGIN_PROFILE_NAME "profile.plugin"
-
#define CONFIG_SECTION_PROFILE "plugin:profile"
class Generator {
diff --git a/collectors/python.d.plugin/adaptec_raid/README.md b/src/collectors/python.d.plugin/adaptec_raid/README.md
index 97a103eb9..97a103eb9 120000
--- a/collectors/python.d.plugin/adaptec_raid/README.md
+++ b/src/collectors/python.d.plugin/adaptec_raid/README.md
diff --git a/collectors/python.d.plugin/adaptec_raid/adaptec_raid.chart.py b/src/collectors/python.d.plugin/adaptec_raid/adaptec_raid.chart.py
index 1995ad681..1995ad681 100644
--- a/collectors/python.d.plugin/adaptec_raid/adaptec_raid.chart.py
+++ b/src/collectors/python.d.plugin/adaptec_raid/adaptec_raid.chart.py
diff --git a/collectors/python.d.plugin/adaptec_raid/adaptec_raid.conf b/src/collectors/python.d.plugin/adaptec_raid/adaptec_raid.conf
index fa462ec83..fa462ec83 100644
--- a/collectors/python.d.plugin/adaptec_raid/adaptec_raid.conf
+++ b/src/collectors/python.d.plugin/adaptec_raid/adaptec_raid.conf
diff --git a/collectors/python.d.plugin/adaptec_raid/integrations/adaptecraid.md b/src/collectors/python.d.plugin/adaptec_raid/integrations/adaptecraid.md
index 13d22ba54..aa28451df 100644
--- a/collectors/python.d.plugin/adaptec_raid/integrations/adaptecraid.md
+++ b/src/collectors/python.d.plugin/adaptec_raid/integrations/adaptecraid.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/adaptec_raid/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/adaptec_raid/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/adaptec_raid/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/adaptec_raid/metadata.yaml"
sidebar_label: "AdaptecRAID"
learn_status: "Published"
-learn_rel_path: "Data Collection/Storage, Mount Points and Filesystems"
+learn_rel_path: "Collecting Metrics/Storage, Mount Points and Filesystems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -84,8 +84,8 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ adaptec_raid_ld_status ](https://github.com/netdata/netdata/blob/master/health/health.d/adaptec_raid.conf) | adaptec_raid.ld_status | logical device status is failed or degraded |
-| [ adaptec_raid_pd_state ](https://github.com/netdata/netdata/blob/master/health/health.d/adaptec_raid.conf) | adaptec_raid.pd_state | physical device state is not online |
+| [ adaptec_raid_ld_status ](https://github.com/netdata/netdata/blob/master/src/health/health.d/adaptec_raid.conf) | adaptec_raid.ld_status | logical device status is failed or degraded |
+| [ adaptec_raid_pd_state ](https://github.com/netdata/netdata/blob/master/src/health/health.d/adaptec_raid.conf) | adaptec_raid.pd_state | physical device state is not online |
## Setup
@@ -127,7 +127,7 @@ The configuration file name for this integration is `python.d/adaptec_raid.conf`
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/alarms/README.md b/src/collectors/python.d.plugin/alarms/README.md
index 85759ae6c..85759ae6c 120000
--- a/collectors/python.d.plugin/alarms/README.md
+++ b/src/collectors/python.d.plugin/alarms/README.md
diff --git a/collectors/python.d.plugin/alarms/alarms.chart.py b/src/collectors/python.d.plugin/alarms/alarms.chart.py
index d19427358..d19427358 100644
--- a/collectors/python.d.plugin/alarms/alarms.chart.py
+++ b/src/collectors/python.d.plugin/alarms/alarms.chart.py
diff --git a/collectors/python.d.plugin/alarms/alarms.conf b/src/collectors/python.d.plugin/alarms/alarms.conf
index 06d76c3b3..06d76c3b3 100644
--- a/collectors/python.d.plugin/alarms/alarms.conf
+++ b/src/collectors/python.d.plugin/alarms/alarms.conf
diff --git a/collectors/python.d.plugin/alarms/integrations/netdata_agent_alarms.md b/src/collectors/python.d.plugin/alarms/integrations/netdata_agent_alarms.md
index 9fb69878a..ca57b00ad 100644
--- a/collectors/python.d.plugin/alarms/integrations/netdata_agent_alarms.md
+++ b/src/collectors/python.d.plugin/alarms/integrations/netdata_agent_alarms.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/alarms/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/alarms/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/alarms/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/alarms/metadata.yaml"
sidebar_label: "Netdata Agent alarms"
learn_status: "Published"
-learn_rel_path: "Data Collection/Other"
+learn_rel_path: "Collecting Metrics/Other"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -87,7 +87,7 @@ The configuration file name for this integration is `python.d/alarms.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/am2320/README.md b/src/collectors/python.d.plugin/am2320/README.md
index 0bc5ea90e..0bc5ea90e 120000
--- a/collectors/python.d.plugin/am2320/README.md
+++ b/src/collectors/python.d.plugin/am2320/README.md
diff --git a/collectors/python.d.plugin/am2320/am2320.chart.py b/src/collectors/python.d.plugin/am2320/am2320.chart.py
index 8e66544bd..8e66544bd 100644
--- a/collectors/python.d.plugin/am2320/am2320.chart.py
+++ b/src/collectors/python.d.plugin/am2320/am2320.chart.py
diff --git a/collectors/python.d.plugin/am2320/am2320.conf b/src/collectors/python.d.plugin/am2320/am2320.conf
index c6b9885fc..c6b9885fc 100644
--- a/collectors/python.d.plugin/am2320/am2320.conf
+++ b/src/collectors/python.d.plugin/am2320/am2320.conf
diff --git a/collectors/python.d.plugin/am2320/integrations/am2320.md b/src/collectors/python.d.plugin/am2320/integrations/am2320.md
index 72b351eb5..db974eb63 100644
--- a/collectors/python.d.plugin/am2320/integrations/am2320.md
+++ b/src/collectors/python.d.plugin/am2320/integrations/am2320.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/am2320/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/am2320/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/am2320/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/am2320/metadata.yaml"
sidebar_label: "AM2320"
learn_status: "Published"
-learn_rel_path: "Data Collection/Hardware Devices and Sensors"
+learn_rel_path: "Collecting Metrics/Hardware Devices and Sensors"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -107,7 +107,7 @@ The configuration file name for this integration is `python.d/am2320.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/am2320/metadata.yaml b/src/collectors/python.d.plugin/am2320/metadata.yaml
index c85cd5f22..c85cd5f22 100644
--- a/collectors/python.d.plugin/am2320/metadata.yaml
+++ b/src/collectors/python.d.plugin/am2320/metadata.yaml
diff --git a/collectors/python.d.plugin/anomalies/anomalies.chart.py b/src/collectors/python.d.plugin/anomalies/anomalies.chart.py
index 24e84cc15..24e84cc15 100644
--- a/collectors/python.d.plugin/anomalies/anomalies.chart.py
+++ b/src/collectors/python.d.plugin/anomalies/anomalies.chart.py
diff --git a/collectors/python.d.plugin/anomalies/anomalies.conf b/src/collectors/python.d.plugin/anomalies/anomalies.conf
index ef867709a..ef867709a 100644
--- a/collectors/python.d.plugin/anomalies/anomalies.conf
+++ b/src/collectors/python.d.plugin/anomalies/anomalies.conf
diff --git a/collectors/python.d.plugin/beanstalk/README.md b/src/collectors/python.d.plugin/beanstalk/README.md
index 4efe13889..4efe13889 120000
--- a/collectors/python.d.plugin/beanstalk/README.md
+++ b/src/collectors/python.d.plugin/beanstalk/README.md
diff --git a/collectors/python.d.plugin/beanstalk/beanstalk.chart.py b/src/collectors/python.d.plugin/beanstalk/beanstalk.chart.py
index 396543e5a..396543e5a 100644
--- a/collectors/python.d.plugin/beanstalk/beanstalk.chart.py
+++ b/src/collectors/python.d.plugin/beanstalk/beanstalk.chart.py
diff --git a/collectors/python.d.plugin/beanstalk/beanstalk.conf b/src/collectors/python.d.plugin/beanstalk/beanstalk.conf
index 6d9773a19..6d9773a19 100644
--- a/collectors/python.d.plugin/beanstalk/beanstalk.conf
+++ b/src/collectors/python.d.plugin/beanstalk/beanstalk.conf
diff --git a/collectors/python.d.plugin/beanstalk/integrations/beanstalk.md b/src/collectors/python.d.plugin/beanstalk/integrations/beanstalk.md
index 5095c0c28..eca485522 100644
--- a/collectors/python.d.plugin/beanstalk/integrations/beanstalk.md
+++ b/src/collectors/python.d.plugin/beanstalk/integrations/beanstalk.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/beanstalk/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/beanstalk/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/beanstalk/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/beanstalk/metadata.yaml"
sidebar_label: "Beanstalk"
learn_status: "Published"
-learn_rel_path: "Data Collection/Message Brokers"
+learn_rel_path: "Collecting Metrics/Message Brokers"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -98,7 +98,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ beanstalk_server_buried_jobs ](https://github.com/netdata/netdata/blob/master/health/health.d/beanstalkd.conf) | beanstalk.current_jobs | number of buried jobs across all tubes. You need to manually kick them so they can be processed. Presence of buried jobs in a tube does not affect new jobs. |
+| [ beanstalk_server_buried_jobs ](https://github.com/netdata/netdata/blob/master/src/health/health.d/beanstalkd.conf) | beanstalk.current_jobs | number of buried jobs across all tubes. You need to manually kick them so they can be processed. Presence of buried jobs in a tube does not affect new jobs. |
## Setup
@@ -118,7 +118,7 @@ The configuration file name for this integration is `python.d/beanstalk.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/bind_rndc/README.md b/src/collectors/python.d.plugin/bind_rndc/README.md
index 03a182ae8..03a182ae8 120000
--- a/collectors/python.d.plugin/bind_rndc/README.md
+++ b/src/collectors/python.d.plugin/bind_rndc/README.md
diff --git a/collectors/python.d.plugin/bind_rndc/bind_rndc.chart.py b/src/collectors/python.d.plugin/bind_rndc/bind_rndc.chart.py
index 9d6c9fec7..9d6c9fec7 100644
--- a/collectors/python.d.plugin/bind_rndc/bind_rndc.chart.py
+++ b/src/collectors/python.d.plugin/bind_rndc/bind_rndc.chart.py
diff --git a/collectors/python.d.plugin/bind_rndc/bind_rndc.conf b/src/collectors/python.d.plugin/bind_rndc/bind_rndc.conf
index 84eaf0594..84eaf0594 100644
--- a/collectors/python.d.plugin/bind_rndc/bind_rndc.conf
+++ b/src/collectors/python.d.plugin/bind_rndc/bind_rndc.conf
diff --git a/collectors/python.d.plugin/bind_rndc/integrations/isc_bind_rndc.md b/src/collectors/python.d.plugin/bind_rndc/integrations/isc_bind_rndc.md
index 163f8282c..0607e47da 100644
--- a/collectors/python.d.plugin/bind_rndc/integrations/isc_bind_rndc.md
+++ b/src/collectors/python.d.plugin/bind_rndc/integrations/isc_bind_rndc.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/bind_rndc/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/bind_rndc/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/bind_rndc/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/bind_rndc/metadata.yaml"
sidebar_label: "ISC Bind (RNDC)"
learn_status: "Published"
-learn_rel_path: "Data Collection/DNS and DHCP Servers"
+learn_rel_path: "Collecting Metrics/DNS and DHCP Servers"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -77,7 +77,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ bind_rndc_stats_file_size ](https://github.com/netdata/netdata/blob/master/health/health.d/bind_rndc.conf) | bind_rndc.stats_size | BIND statistics-file size |
+| [ bind_rndc_stats_file_size ](https://github.com/netdata/netdata/blob/master/src/health/health.d/bind_rndc.conf) | bind_rndc.stats_size | BIND statistics-file size |
## Setup
@@ -140,7 +140,7 @@ The configuration file name for this integration is `python.d/bind_rndc.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/boinc/README.md b/src/collectors/python.d.plugin/boinc/README.md
index 22c10ca17..22c10ca17 120000
--- a/collectors/python.d.plugin/boinc/README.md
+++ b/src/collectors/python.d.plugin/boinc/README.md
diff --git a/collectors/python.d.plugin/boinc/boinc.chart.py b/src/collectors/python.d.plugin/boinc/boinc.chart.py
index a31eda1c2..a31eda1c2 100644
--- a/collectors/python.d.plugin/boinc/boinc.chart.py
+++ b/src/collectors/python.d.plugin/boinc/boinc.chart.py
diff --git a/collectors/python.d.plugin/boinc/boinc.conf b/src/collectors/python.d.plugin/boinc/boinc.conf
index 16edf55c4..16edf55c4 100644
--- a/collectors/python.d.plugin/boinc/boinc.conf
+++ b/src/collectors/python.d.plugin/boinc/boinc.conf
diff --git a/collectors/python.d.plugin/boinc/integrations/boinc.md b/src/collectors/python.d.plugin/boinc/integrations/boinc.md
index d6874d455..bfd1f7198 100644
--- a/collectors/python.d.plugin/boinc/integrations/boinc.md
+++ b/src/collectors/python.d.plugin/boinc/integrations/boinc.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/boinc/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/boinc/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/boinc/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/boinc/metadata.yaml"
sidebar_label: "BOINC"
learn_status: "Published"
-learn_rel_path: "Data Collection/Distributed Computing Systems"
+learn_rel_path: "Collecting Metrics/Distributed Computing Systems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -77,10 +77,10 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ boinc_total_tasks ](https://github.com/netdata/netdata/blob/master/health/health.d/boinc.conf) | boinc.tasks | average number of total tasks over the last 10 minutes |
-| [ boinc_active_tasks ](https://github.com/netdata/netdata/blob/master/health/health.d/boinc.conf) | boinc.tasks | average number of active tasks over the last 10 minutes |
-| [ boinc_compute_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/boinc.conf) | boinc.states | average number of compute errors over the last 10 minutes |
-| [ boinc_upload_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/boinc.conf) | boinc.states | average number of failed uploads over the last 10 minutes |
+| [ boinc_total_tasks ](https://github.com/netdata/netdata/blob/master/src/health/health.d/boinc.conf) | boinc.tasks | average number of total tasks over the last 10 minutes |
+| [ boinc_active_tasks ](https://github.com/netdata/netdata/blob/master/src/health/health.d/boinc.conf) | boinc.tasks | average number of active tasks over the last 10 minutes |
+| [ boinc_compute_errors ](https://github.com/netdata/netdata/blob/master/src/health/health.d/boinc.conf) | boinc.states | average number of compute errors over the last 10 minutes |
+| [ boinc_upload_errors ](https://github.com/netdata/netdata/blob/master/src/health/health.d/boinc.conf) | boinc.states | average number of failed uploads over the last 10 minutes |
## Setup
@@ -100,7 +100,7 @@ The configuration file name for this integration is `python.d/boinc.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/ceph/README.md b/src/collectors/python.d.plugin/ceph/README.md
index 654248b70..654248b70 120000
--- a/collectors/python.d.plugin/ceph/README.md
+++ b/src/collectors/python.d.plugin/ceph/README.md
diff --git a/collectors/python.d.plugin/ceph/ceph.chart.py b/src/collectors/python.d.plugin/ceph/ceph.chart.py
index 4bcbe1979..4bcbe1979 100644
--- a/collectors/python.d.plugin/ceph/ceph.chart.py
+++ b/src/collectors/python.d.plugin/ceph/ceph.chart.py
diff --git a/collectors/python.d.plugin/ceph/ceph.conf b/src/collectors/python.d.plugin/ceph/ceph.conf
index 81788e866..81788e866 100644
--- a/collectors/python.d.plugin/ceph/ceph.conf
+++ b/src/collectors/python.d.plugin/ceph/ceph.conf
diff --git a/collectors/python.d.plugin/ceph/integrations/ceph.md b/src/collectors/python.d.plugin/ceph/integrations/ceph.md
index cfda01fbe..d0bc02694 100644
--- a/collectors/python.d.plugin/ceph/integrations/ceph.md
+++ b/src/collectors/python.d.plugin/ceph/integrations/ceph.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/ceph/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/ceph/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/ceph/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/ceph/metadata.yaml"
sidebar_label: "Ceph"
learn_status: "Published"
-learn_rel_path: "Data Collection/Storage, Mount Points and Filesystems"
+learn_rel_path: "Collecting Metrics/Storage, Mount Points and Filesystems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -88,7 +88,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ ceph_cluster_space_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/ceph.conf) | ceph.general_usage | cluster disk space utilization |
+| [ ceph_cluster_space_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ceph.conf) | ceph.general_usage | cluster disk space utilization |
## Setup
@@ -116,7 +116,7 @@ The configuration file name for this integration is `python.d/ceph.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/changefinder/README.md b/src/collectors/python.d.plugin/changefinder/README.md
index 0ca704eb1..0ca704eb1 120000
--- a/collectors/python.d.plugin/changefinder/README.md
+++ b/src/collectors/python.d.plugin/changefinder/README.md
diff --git a/collectors/python.d.plugin/changefinder/changefinder.chart.py b/src/collectors/python.d.plugin/changefinder/changefinder.chart.py
index 2a69cd9f5..2a69cd9f5 100644
--- a/collectors/python.d.plugin/changefinder/changefinder.chart.py
+++ b/src/collectors/python.d.plugin/changefinder/changefinder.chart.py
diff --git a/collectors/python.d.plugin/changefinder/changefinder.conf b/src/collectors/python.d.plugin/changefinder/changefinder.conf
index 56a681f1e..56a681f1e 100644
--- a/collectors/python.d.plugin/changefinder/changefinder.conf
+++ b/src/collectors/python.d.plugin/changefinder/changefinder.conf
diff --git a/collectors/python.d.plugin/changefinder/integrations/python.d_changefinder.md b/src/collectors/python.d.plugin/changefinder/integrations/python.d_changefinder.md
index c338c9374..894e8c41d 100644
--- a/collectors/python.d.plugin/changefinder/integrations/python.d_changefinder.md
+++ b/src/collectors/python.d.plugin/changefinder/integrations/python.d_changefinder.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/changefinder/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/changefinder/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/changefinder/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/changefinder/metadata.yaml"
sidebar_label: "python.d changefinder"
learn_status: "Published"
-learn_rel_path: "Data Collection/Other"
+learn_rel_path: "Collecting Metrics/Other"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -121,7 +121,7 @@ The configuration file name for this integration is `python.d/changefinder.conf`
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/changefinder/metadata.yaml b/src/collectors/python.d.plugin/changefinder/metadata.yaml
index 170d9146a..170d9146a 100644
--- a/collectors/python.d.plugin/changefinder/metadata.yaml
+++ b/src/collectors/python.d.plugin/changefinder/metadata.yaml
diff --git a/collectors/python.d.plugin/dovecot/README.md b/src/collectors/python.d.plugin/dovecot/README.md
index c4749cedc..c4749cedc 120000
--- a/collectors/python.d.plugin/dovecot/README.md
+++ b/src/collectors/python.d.plugin/dovecot/README.md
diff --git a/collectors/python.d.plugin/dovecot/dovecot.chart.py b/src/collectors/python.d.plugin/dovecot/dovecot.chart.py
index dfaef28b5..dfaef28b5 100644
--- a/collectors/python.d.plugin/dovecot/dovecot.chart.py
+++ b/src/collectors/python.d.plugin/dovecot/dovecot.chart.py
diff --git a/collectors/python.d.plugin/dovecot/dovecot.conf b/src/collectors/python.d.plugin/dovecot/dovecot.conf
index 451dbc9ac..451dbc9ac 100644
--- a/collectors/python.d.plugin/dovecot/dovecot.conf
+++ b/src/collectors/python.d.plugin/dovecot/dovecot.conf
diff --git a/collectors/python.d.plugin/dovecot/integrations/dovecot.md b/src/collectors/python.d.plugin/dovecot/integrations/dovecot.md
index 4e7952765..be515ed9f 100644
--- a/collectors/python.d.plugin/dovecot/integrations/dovecot.md
+++ b/src/collectors/python.d.plugin/dovecot/integrations/dovecot.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/dovecot/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/dovecot/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/dovecot/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/dovecot/metadata.yaml"
sidebar_label: "Dovecot"
learn_status: "Published"
-learn_rel_path: "Data Collection/Mail Servers"
+learn_rel_path: "Collecting Metrics/Mail Servers"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -100,7 +100,7 @@ The configuration file name for this integration is `python.d/dovecot.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/dovecot/metadata.yaml b/src/collectors/python.d.plugin/dovecot/metadata.yaml
index b247da846..b247da846 100644
--- a/collectors/python.d.plugin/dovecot/metadata.yaml
+++ b/src/collectors/python.d.plugin/dovecot/metadata.yaml
diff --git a/collectors/python.d.plugin/example/README.md b/src/collectors/python.d.plugin/example/README.md
index 55877a99a..55877a99a 120000
--- a/collectors/python.d.plugin/example/README.md
+++ b/src/collectors/python.d.plugin/example/README.md
diff --git a/collectors/python.d.plugin/example/example.chart.py b/src/collectors/python.d.plugin/example/example.chart.py
index d6c0b6658..d6c0b6658 100644
--- a/collectors/python.d.plugin/example/example.chart.py
+++ b/src/collectors/python.d.plugin/example/example.chart.py
diff --git a/collectors/python.d.plugin/example/example.conf b/src/collectors/python.d.plugin/example/example.conf
index 31261b840..31261b840 100644
--- a/collectors/python.d.plugin/example/example.conf
+++ b/src/collectors/python.d.plugin/example/example.conf
diff --git a/collectors/python.d.plugin/example/integrations/example_collector.md b/src/collectors/python.d.plugin/example/integrations/example_collector.md
index 7dded67ba..3d031a1a4 100644
--- a/collectors/python.d.plugin/example/integrations/example_collector.md
+++ b/src/collectors/python.d.plugin/example/integrations/example_collector.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/example/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/example/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/example/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/example/metadata.yaml"
sidebar_label: "Example collector"
learn_status: "Published"
-learn_rel_path: "Data Collection/Other"
+learn_rel_path: "Collecting Metrics/Other"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -19,7 +19,7 @@ Module: example
Example collector that generates some random numbers as metrics.
-If you want to write your own collector, read our [writing a new Python module](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/README.md#how-to-write-a-new-module) tutorial.
+If you want to write your own collector, read our [writing a new Python module](https://github.com/netdata/netdata/blob/master/src/collectors/python.d.plugin/README.md#how-to-write-a-new-module) tutorial.
The `get_data()` function uses `random.randint()` to generate a random number which will be collected as a metric.
@@ -87,7 +87,7 @@ The configuration file name for this integration is `python.d/example.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/exim/README.md b/src/collectors/python.d.plugin/exim/README.md
index f1f2ef9f9..f1f2ef9f9 120000
--- a/collectors/python.d.plugin/exim/README.md
+++ b/src/collectors/python.d.plugin/exim/README.md
diff --git a/collectors/python.d.plugin/exim/exim.chart.py b/src/collectors/python.d.plugin/exim/exim.chart.py
index 7238a1bea..7238a1bea 100644
--- a/collectors/python.d.plugin/exim/exim.chart.py
+++ b/src/collectors/python.d.plugin/exim/exim.chart.py
diff --git a/collectors/python.d.plugin/exim/exim.conf b/src/collectors/python.d.plugin/exim/exim.conf
index 3b7e65922..3b7e65922 100644
--- a/collectors/python.d.plugin/exim/exim.conf
+++ b/src/collectors/python.d.plugin/exim/exim.conf
diff --git a/collectors/python.d.plugin/exim/integrations/exim.md b/src/collectors/python.d.plugin/exim/integrations/exim.md
index f0ae33d3e..0e77fbeb6 100644
--- a/collectors/python.d.plugin/exim/integrations/exim.md
+++ b/src/collectors/python.d.plugin/exim/integrations/exim.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/exim/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/exim/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/exim/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/exim/metadata.yaml"
sidebar_label: "Exim"
learn_status: "Published"
-learn_rel_path: "Data Collection/Mail Servers"
+learn_rel_path: "Collecting Metrics/Mail Servers"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -106,7 +106,7 @@ The configuration file name for this integration is `python.d/exim.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/exim/metadata.yaml b/src/collectors/python.d.plugin/exim/metadata.yaml
index a8be02d99..a8be02d99 100644
--- a/collectors/python.d.plugin/exim/metadata.yaml
+++ b/src/collectors/python.d.plugin/exim/metadata.yaml
diff --git a/collectors/python.d.plugin/fail2ban/README.md b/src/collectors/python.d.plugin/fail2ban/README.md
index 642a8bcf5..642a8bcf5 120000
--- a/collectors/python.d.plugin/fail2ban/README.md
+++ b/src/collectors/python.d.plugin/fail2ban/README.md
diff --git a/collectors/python.d.plugin/fail2ban/fail2ban.chart.py b/src/collectors/python.d.plugin/fail2ban/fail2ban.chart.py
index 76f6d92b4..76f6d92b4 100644
--- a/collectors/python.d.plugin/fail2ban/fail2ban.chart.py
+++ b/src/collectors/python.d.plugin/fail2ban/fail2ban.chart.py
diff --git a/collectors/python.d.plugin/fail2ban/fail2ban.conf b/src/collectors/python.d.plugin/fail2ban/fail2ban.conf
index a36436b51..a36436b51 100644
--- a/collectors/python.d.plugin/fail2ban/fail2ban.conf
+++ b/src/collectors/python.d.plugin/fail2ban/fail2ban.conf
diff --git a/collectors/python.d.plugin/fail2ban/integrations/fail2ban.md b/src/collectors/python.d.plugin/fail2ban/integrations/fail2ban.md
index a7116be5e..fa1b256ba 100644
--- a/collectors/python.d.plugin/fail2ban/integrations/fail2ban.md
+++ b/src/collectors/python.d.plugin/fail2ban/integrations/fail2ban.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/fail2ban/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/fail2ban/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/fail2ban/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/fail2ban/metadata.yaml"
sidebar_label: "Fail2ban"
learn_status: "Published"
-learn_rel_path: "Data Collection/Authentication and Authorization"
+learn_rel_path: "Collecting Metrics/Authentication and Authorization"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -126,7 +126,7 @@ The configuration file name for this integration is `python.d/fail2ban.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/fail2ban/metadata.yaml b/src/collectors/python.d.plugin/fail2ban/metadata.yaml
index 61f762679..61f762679 100644
--- a/collectors/python.d.plugin/fail2ban/metadata.yaml
+++ b/src/collectors/python.d.plugin/fail2ban/metadata.yaml
diff --git a/collectors/python.d.plugin/gearman/README.md b/src/collectors/python.d.plugin/gearman/README.md
index 70189d698..70189d698 120000
--- a/collectors/python.d.plugin/gearman/README.md
+++ b/src/collectors/python.d.plugin/gearman/README.md
diff --git a/collectors/python.d.plugin/gearman/gearman.chart.py b/src/collectors/python.d.plugin/gearman/gearman.chart.py
index 5e280a4d8..5e280a4d8 100644
--- a/collectors/python.d.plugin/gearman/gearman.chart.py
+++ b/src/collectors/python.d.plugin/gearman/gearman.chart.py
diff --git a/collectors/python.d.plugin/gearman/gearman.conf b/src/collectors/python.d.plugin/gearman/gearman.conf
index 635e893ef..635e893ef 100644
--- a/collectors/python.d.plugin/gearman/gearman.conf
+++ b/src/collectors/python.d.plugin/gearman/gearman.conf
diff --git a/collectors/python.d.plugin/gearman/integrations/gearman.md b/src/collectors/python.d.plugin/gearman/integrations/gearman.md
index 3923d1401..c9612c041 100644
--- a/collectors/python.d.plugin/gearman/integrations/gearman.md
+++ b/src/collectors/python.d.plugin/gearman/integrations/gearman.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/gearman/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/gearman/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/gearman/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/gearman/metadata.yaml"
sidebar_label: "Gearman"
learn_status: "Published"
-learn_rel_path: "Data Collection/Distributed Computing Systems"
+learn_rel_path: "Collecting Metrics/Distributed Computing Systems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -86,7 +86,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ gearman_workers_queued ](https://github.com/netdata/netdata/blob/master/health/health.d/gearman.conf) | gearman.single_job | average number of queued jobs over the last 10 minutes |
+| [ gearman_workers_queued ](https://github.com/netdata/netdata/blob/master/src/health/health.d/gearman.conf) | gearman.single_job | average number of queued jobs over the last 10 minutes |
## Setup
@@ -106,7 +106,7 @@ The configuration file name for this integration is `python.d/gearman.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/go_expvar/README.md b/src/collectors/python.d.plugin/go_expvar/README.md
index f28a82f34..f28a82f34 120000
--- a/collectors/python.d.plugin/go_expvar/README.md
+++ b/src/collectors/python.d.plugin/go_expvar/README.md
diff --git a/collectors/python.d.plugin/go_expvar/go_expvar.chart.py b/src/collectors/python.d.plugin/go_expvar/go_expvar.chart.py
index dca010817..dca010817 100644
--- a/collectors/python.d.plugin/go_expvar/go_expvar.chart.py
+++ b/src/collectors/python.d.plugin/go_expvar/go_expvar.chart.py
diff --git a/collectors/python.d.plugin/go_expvar/go_expvar.conf b/src/collectors/python.d.plugin/go_expvar/go_expvar.conf
index 4b821cde9..088fca9cf 100644
--- a/collectors/python.d.plugin/go_expvar/go_expvar.conf
+++ b/src/collectors/python.d.plugin/go_expvar/go_expvar.conf
@@ -74,7 +74,7 @@
#
# Please visit the module wiki page for more information on how to use the extra_charts variable:
#
-# https://github.com/netdata/netdata/tree/master/collectors/python.d.plugin/go_expvar
+# https://github.com/netdata/netdata/tree/master/src/collectors/python.d.plugin/go_expvar
#
# Configuration example
# ---------------------
diff --git a/collectors/python.d.plugin/go_expvar/integrations/go_applications_expvar.md b/src/collectors/python.d.plugin/go_expvar/integrations/go_applications_expvar.md
index 8d61fa2ae..b34c36541 100644
--- a/collectors/python.d.plugin/go_expvar/integrations/go_applications_expvar.md
+++ b/src/collectors/python.d.plugin/go_expvar/integrations/go_applications_expvar.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/go_expvar/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/go_expvar/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/go_expvar/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/go_expvar/metadata.yaml"
sidebar_label: "Go applications (EXPVAR)"
learn_status: "Published"
-learn_rel_path: "Data Collection/APM"
+learn_rel_path: "Collecting Metrics/APM"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -84,14 +84,14 @@ There are no alerts configured by default for this integration.
#### Enable the go_expvar collector
-The `go_expvar` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`, to edit the `python.d.conf` file.
+The `go_expvar` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md), which is typically at `/etc/netdata`, to edit the `python.d.conf` file.
```bash
cd /etc/netdata # Replace this path with your Netdata config directory, if different
sudo ./edit-config python.d.conf
```
-Change the value of the `go_expvar` setting to `yes`. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
+Change the value of the `go_expvar` setting to `yes`. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/packaging/installer/README.md#maintaining-a-netdata-agent-installation) for your system.
#### Sample `expvar` usage in a Go application
@@ -172,7 +172,7 @@ The configuration file name for this integration is `python.d/go_expvar.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
@@ -231,8 +231,8 @@ See [this issue](https://github.com/netdata/netdata/pull/1902#issuecomment-28449
Please see these two links to the official Netdata documentation for more information about the values:
-- [External plugins - charts](https://github.com/netdata/netdata/blob/master/collectors/plugins.d/README.md#chart)
-- [Chart variables](https://github.com/netdata/netdata/blob/master/collectors/python.d.plugin/README.md#global-variables-order-and-chart)
+- [External plugins - charts](https://github.com/netdata/netdata/blob/master/src/collectors/plugins.d/README.md#chart)
+- [Chart variables](https://github.com/netdata/netdata/blob/master/src/collectors/python.d.plugin/README.md#global-variables-order-and-chart)
**Line definitions**
@@ -255,7 +255,7 @@ hidden: False
```
Please see the following link for more information about the options and their default values:
-[External plugins - dimensions](https://github.com/netdata/netdata/blob/master/collectors/plugins.d/README.md#dimension)
+[External plugins - dimensions](https://github.com/netdata/netdata/blob/master/src/collectors/plugins.d/README.md#dimension)
Apart from top-level expvars, this plugin can also parse expvars stored in a multi-level map;
All dicts in the resulting JSON document are then flattened to one level.
diff --git a/collectors/python.d.plugin/haproxy/haproxy.chart.py b/src/collectors/python.d.plugin/haproxy/haproxy.chart.py
index f412febb7..f412febb7 100644
--- a/collectors/python.d.plugin/haproxy/haproxy.chart.py
+++ b/src/collectors/python.d.plugin/haproxy/haproxy.chart.py
diff --git a/collectors/python.d.plugin/haproxy/haproxy.conf b/src/collectors/python.d.plugin/haproxy/haproxy.conf
index 10a0df3c3..10a0df3c3 100644
--- a/collectors/python.d.plugin/haproxy/haproxy.conf
+++ b/src/collectors/python.d.plugin/haproxy/haproxy.conf
diff --git a/collectors/python.d.plugin/hddtemp/README.md b/src/collectors/python.d.plugin/hddtemp/README.md
index 95c7593f8..95c7593f8 120000
--- a/collectors/python.d.plugin/hddtemp/README.md
+++ b/src/collectors/python.d.plugin/hddtemp/README.md
diff --git a/collectors/python.d.plugin/hddtemp/hddtemp.chart.py b/src/collectors/python.d.plugin/hddtemp/hddtemp.chart.py
index 6427aa180..6427aa180 100644
--- a/collectors/python.d.plugin/hddtemp/hddtemp.chart.py
+++ b/src/collectors/python.d.plugin/hddtemp/hddtemp.chart.py
diff --git a/collectors/python.d.plugin/hddtemp/hddtemp.conf b/src/collectors/python.d.plugin/hddtemp/hddtemp.conf
index b2d7aef63..b2d7aef63 100644
--- a/collectors/python.d.plugin/hddtemp/hddtemp.conf
+++ b/src/collectors/python.d.plugin/hddtemp/hddtemp.conf
diff --git a/collectors/python.d.plugin/hddtemp/integrations/hdd_temperature.md b/src/collectors/python.d.plugin/hddtemp/integrations/hdd_temperature.md
index 4a1504f07..9b39d1371 100644
--- a/collectors/python.d.plugin/hddtemp/integrations/hdd_temperature.md
+++ b/src/collectors/python.d.plugin/hddtemp/integrations/hdd_temperature.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/hddtemp/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/hddtemp/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/hddtemp/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/hddtemp/metadata.yaml"
sidebar_label: "HDD temperature"
learn_status: "Published"
-learn_rel_path: "Data Collection/Hardware Devices and Sensors"
+learn_rel_path: "Collecting Metrics/Hardware Devices and Sensors"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -96,7 +96,7 @@ The configuration file name for this integration is `python.d/hddtemp.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/hddtemp/metadata.yaml b/src/collectors/python.d.plugin/hddtemp/metadata.yaml
index d8b56fc66..d8b56fc66 100644
--- a/collectors/python.d.plugin/hddtemp/metadata.yaml
+++ b/src/collectors/python.d.plugin/hddtemp/metadata.yaml
diff --git a/collectors/python.d.plugin/hpssa/README.md b/src/collectors/python.d.plugin/hpssa/README.md
index 82802d8b4..82802d8b4 120000
--- a/collectors/python.d.plugin/hpssa/README.md
+++ b/src/collectors/python.d.plugin/hpssa/README.md
diff --git a/collectors/python.d.plugin/hpssa/hpssa.chart.py b/src/collectors/python.d.plugin/hpssa/hpssa.chart.py
index 66be00837..66be00837 100644
--- a/collectors/python.d.plugin/hpssa/hpssa.chart.py
+++ b/src/collectors/python.d.plugin/hpssa/hpssa.chart.py
diff --git a/collectors/python.d.plugin/hpssa/hpssa.conf b/src/collectors/python.d.plugin/hpssa/hpssa.conf
index cc50c9836..cc50c9836 100644
--- a/collectors/python.d.plugin/hpssa/hpssa.conf
+++ b/src/collectors/python.d.plugin/hpssa/hpssa.conf
diff --git a/collectors/python.d.plugin/hpssa/integrations/hp_smart_storage_arrays.md b/src/collectors/python.d.plugin/hpssa/integrations/hp_smart_storage_arrays.md
index d46cc9065..136ec0190 100644
--- a/collectors/python.d.plugin/hpssa/integrations/hp_smart_storage_arrays.md
+++ b/src/collectors/python.d.plugin/hpssa/integrations/hp_smart_storage_arrays.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/hpssa/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/hpssa/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/hpssa/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/hpssa/metadata.yaml"
sidebar_label: "HP Smart Storage Arrays"
learn_status: "Published"
-learn_rel_path: "Data Collection/Storage, Mount Points and Filesystems"
+learn_rel_path: "Collecting Metrics/Storage, Mount Points and Filesystems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -82,14 +82,14 @@ There are no alerts configured by default for this integration.
#### Enable the hpssa collector
-The `hpssa` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`, to edit the `python.d.conf` file.
+The `hpssa` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md), which is typically at `/etc/netdata`, to edit the `python.d.conf` file.
```bash
cd /etc/netdata # Replace this path with your Netdata config directory, if different
sudo ./edit-config python.d.conf
```
-Change the value of the `hpssa` setting to `yes`. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
+Change the value of the `hpssa` setting to `yes`. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/packaging/installer/README.md#maintaining-a-netdata-agent-installation) for your system.
#### Allow user netdata to execute `ssacli` as root.
@@ -129,7 +129,7 @@ The configuration file name for this integration is `python.d/hpssa.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/icecast/README.md b/src/collectors/python.d.plugin/icecast/README.md
index db3c1b572..db3c1b572 120000
--- a/collectors/python.d.plugin/icecast/README.md
+++ b/src/collectors/python.d.plugin/icecast/README.md
diff --git a/collectors/python.d.plugin/icecast/icecast.chart.py b/src/collectors/python.d.plugin/icecast/icecast.chart.py
index a967d1779..a967d1779 100644
--- a/collectors/python.d.plugin/icecast/icecast.chart.py
+++ b/src/collectors/python.d.plugin/icecast/icecast.chart.py
diff --git a/collectors/python.d.plugin/icecast/icecast.conf b/src/collectors/python.d.plugin/icecast/icecast.conf
index a33074aef..a33074aef 100644
--- a/collectors/python.d.plugin/icecast/icecast.conf
+++ b/src/collectors/python.d.plugin/icecast/icecast.conf
diff --git a/collectors/python.d.plugin/icecast/integrations/icecast.md b/src/collectors/python.d.plugin/icecast/integrations/icecast.md
index 12d7d59ee..751bb0e30 100644
--- a/collectors/python.d.plugin/icecast/integrations/icecast.md
+++ b/src/collectors/python.d.plugin/icecast/integrations/icecast.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/icecast/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/icecast/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/icecast/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/icecast/metadata.yaml"
sidebar_label: "Icecast"
learn_status: "Published"
-learn_rel_path: "Data Collection/Media Services"
+learn_rel_path: "Collecting Metrics/Media Services"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -89,7 +89,7 @@ The configuration file name for this integration is `python.d/icecast.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/icecast/metadata.yaml b/src/collectors/python.d.plugin/icecast/metadata.yaml
index 4bcf5e39f..4bcf5e39f 100644
--- a/collectors/python.d.plugin/icecast/metadata.yaml
+++ b/src/collectors/python.d.plugin/icecast/metadata.yaml
diff --git a/collectors/python.d.plugin/ipfs/README.md b/src/collectors/python.d.plugin/ipfs/README.md
index eee6a07b2..eee6a07b2 120000
--- a/collectors/python.d.plugin/ipfs/README.md
+++ b/src/collectors/python.d.plugin/ipfs/README.md
diff --git a/collectors/python.d.plugin/ipfs/integrations/ipfs.md b/src/collectors/python.d.plugin/ipfs/integrations/ipfs.md
index 77dc745aa..6b742d4d8 100644
--- a/collectors/python.d.plugin/ipfs/integrations/ipfs.md
+++ b/src/collectors/python.d.plugin/ipfs/integrations/ipfs.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/ipfs/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/ipfs/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/ipfs/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/ipfs/metadata.yaml"
sidebar_label: "IPFS"
learn_status: "Published"
-learn_rel_path: "Data Collection/Storage, Mount Points and Filesystems"
+learn_rel_path: "Collecting Metrics/Storage, Mount Points and Filesystems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -81,7 +81,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ ipfs_datastore_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/ipfs.conf) | ipfs.repo_size | IPFS datastore utilization |
+| [ ipfs_datastore_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/ipfs.conf) | ipfs.repo_size | IPFS datastore utilization |
## Setup
@@ -98,7 +98,7 @@ The configuration file name for this integration is `python.d/ipfs.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/ipfs/ipfs.chart.py b/src/collectors/python.d.plugin/ipfs/ipfs.chart.py
index abfc9c492..abfc9c492 100644
--- a/collectors/python.d.plugin/ipfs/ipfs.chart.py
+++ b/src/collectors/python.d.plugin/ipfs/ipfs.chart.py
diff --git a/collectors/python.d.plugin/ipfs/ipfs.conf b/src/collectors/python.d.plugin/ipfs/ipfs.conf
index 8b167b399..8b167b399 100644
--- a/collectors/python.d.plugin/ipfs/ipfs.conf
+++ b/src/collectors/python.d.plugin/ipfs/ipfs.conf
diff --git a/collectors/python.d.plugin/litespeed/README.md b/src/collectors/python.d.plugin/litespeed/README.md
index e7418b3dc..e7418b3dc 120000
--- a/collectors/python.d.plugin/litespeed/README.md
+++ b/src/collectors/python.d.plugin/litespeed/README.md
diff --git a/collectors/python.d.plugin/litespeed/integrations/litespeed.md b/src/collectors/python.d.plugin/litespeed/integrations/litespeed.md
index 87f2d0b12..2755f7448 100644
--- a/collectors/python.d.plugin/litespeed/integrations/litespeed.md
+++ b/src/collectors/python.d.plugin/litespeed/integrations/litespeed.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/litespeed/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/litespeed/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/litespeed/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/litespeed/metadata.yaml"
sidebar_label: "Litespeed"
learn_status: "Published"
-learn_rel_path: "Data Collection/Web Servers and Web Proxies"
+learn_rel_path: "Collecting Metrics/Web Servers and Web Proxies"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -94,7 +94,7 @@ The configuration file name for this integration is `python.d/litespeed.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/litespeed/litespeed.chart.py b/src/collectors/python.d.plugin/litespeed/litespeed.chart.py
index 7ef8189ea..7ef8189ea 100644
--- a/collectors/python.d.plugin/litespeed/litespeed.chart.py
+++ b/src/collectors/python.d.plugin/litespeed/litespeed.chart.py
diff --git a/collectors/python.d.plugin/litespeed/litespeed.conf b/src/collectors/python.d.plugin/litespeed/litespeed.conf
index a326e184e..a326e184e 100644
--- a/collectors/python.d.plugin/litespeed/litespeed.conf
+++ b/src/collectors/python.d.plugin/litespeed/litespeed.conf
diff --git a/collectors/python.d.plugin/litespeed/metadata.yaml b/src/collectors/python.d.plugin/litespeed/metadata.yaml
index 400f3a7f6..400f3a7f6 100644
--- a/collectors/python.d.plugin/litespeed/metadata.yaml
+++ b/src/collectors/python.d.plugin/litespeed/metadata.yaml
diff --git a/collectors/python.d.plugin/megacli/README.md b/src/collectors/python.d.plugin/megacli/README.md
index e5df4d41d..e5df4d41d 120000
--- a/collectors/python.d.plugin/megacli/README.md
+++ b/src/collectors/python.d.plugin/megacli/README.md
diff --git a/collectors/python.d.plugin/megacli/integrations/megacli.md b/src/collectors/python.d.plugin/megacli/integrations/megacli.md
index 0c4af78a9..23f6770cb 100644
--- a/collectors/python.d.plugin/megacli/integrations/megacli.md
+++ b/src/collectors/python.d.plugin/megacli/integrations/megacli.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/megacli/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/megacli/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/megacli/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/megacli/metadata.yaml"
sidebar_label: "MegaCLI"
learn_status: "Published"
-learn_rel_path: "Data Collection/Storage, Mount Points and Filesystems"
+learn_rel_path: "Collecting Metrics/Storage, Mount Points and Filesystems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -96,11 +96,11 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ megacli_adapter_state ](https://github.com/netdata/netdata/blob/master/health/health.d/megacli.conf) | megacli.adapter_degraded | adapter is in the degraded state (0: false, 1: true) |
-| [ megacli_pd_media_errors ](https://github.com/netdata/netdata/blob/master/health/health.d/megacli.conf) | megacli.pd_media_error | number of physical drive media errors |
-| [ megacli_pd_predictive_failures ](https://github.com/netdata/netdata/blob/master/health/health.d/megacli.conf) | megacli.pd_predictive_failure | number of physical drive predictive failures |
-| [ megacli_bbu_relative_charge ](https://github.com/netdata/netdata/blob/master/health/health.d/megacli.conf) | megacli.bbu_relative_charge | average battery backup unit (BBU) relative state of charge over the last 10 seconds |
-| [ megacli_bbu_cycle_count ](https://github.com/netdata/netdata/blob/master/health/health.d/megacli.conf) | megacli.bbu_cycle_count | average battery backup unit (BBU) charge cycles count over the last 10 seconds |
+| [ megacli_adapter_state ](https://github.com/netdata/netdata/blob/master/src/health/health.d/megacli.conf) | megacli.adapter_degraded | adapter is in the degraded state (0: false, 1: true) |
+| [ megacli_pd_media_errors ](https://github.com/netdata/netdata/blob/master/src/health/health.d/megacli.conf) | megacli.pd_media_error | number of physical drive media errors |
+| [ megacli_pd_predictive_failures ](https://github.com/netdata/netdata/blob/master/src/health/health.d/megacli.conf) | megacli.pd_predictive_failure | number of physical drive predictive failures |
+| [ megacli_bbu_relative_charge ](https://github.com/netdata/netdata/blob/master/src/health/health.d/megacli.conf) | megacli.bbu_relative_charge | average battery backup unit (BBU) relative state of charge over the last 10 seconds |
+| [ megacli_bbu_cycle_count ](https://github.com/netdata/netdata/blob/master/src/health/health.d/megacli.conf) | megacli.bbu_cycle_count | average battery backup unit (BBU) charge cycles count over the last 10 seconds |
## Setup
@@ -142,7 +142,7 @@ The configuration file name for this integration is `python.d/megacli.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/megacli/megacli.chart.py b/src/collectors/python.d.plugin/megacli/megacli.chart.py
index 8222092a8..8222092a8 100644
--- a/collectors/python.d.plugin/megacli/megacli.chart.py
+++ b/src/collectors/python.d.plugin/megacli/megacli.chart.py
diff --git a/collectors/python.d.plugin/megacli/megacli.conf b/src/collectors/python.d.plugin/megacli/megacli.conf
index 1af4292d9..1af4292d9 100644
--- a/collectors/python.d.plugin/megacli/megacli.conf
+++ b/src/collectors/python.d.plugin/megacli/megacli.conf
diff --git a/collectors/python.d.plugin/memcached/README.md b/src/collectors/python.d.plugin/memcached/README.md
index 2cb76d33c..2cb76d33c 120000
--- a/collectors/python.d.plugin/memcached/README.md
+++ b/src/collectors/python.d.plugin/memcached/README.md
diff --git a/collectors/python.d.plugin/memcached/integrations/memcached.md b/src/collectors/python.d.plugin/memcached/integrations/memcached.md
index 113b86c8c..a3b957eb9 100644
--- a/collectors/python.d.plugin/memcached/integrations/memcached.md
+++ b/src/collectors/python.d.plugin/memcached/integrations/memcached.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/memcached/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/memcached/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/memcached/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/memcached/metadata.yaml"
sidebar_label: "Memcached"
learn_status: "Published"
-learn_rel_path: "Data Collection/Databases"
+learn_rel_path: "Collecting Metrics/Databases"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -88,9 +88,9 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ memcached_cache_memory_usage ](https://github.com/netdata/netdata/blob/master/health/health.d/memcached.conf) | memcached.cache | cache memory utilization |
-| [ memcached_cache_fill_rate ](https://github.com/netdata/netdata/blob/master/health/health.d/memcached.conf) | memcached.cache | average rate the cache fills up (positive), or frees up (negative) space over the last hour |
-| [ memcached_out_of_cache_space_time ](https://github.com/netdata/netdata/blob/master/health/health.d/memcached.conf) | memcached.cache | estimated time the cache will run out of space if the system continues to add data at the same rate as the past hour |
+| [ memcached_cache_memory_usage ](https://github.com/netdata/netdata/blob/master/src/health/health.d/memcached.conf) | memcached.cache | cache memory utilization |
+| [ memcached_cache_fill_rate ](https://github.com/netdata/netdata/blob/master/src/health/health.d/memcached.conf) | memcached.cache | average rate the cache fills up (positive), or frees up (negative) space over the last hour |
+| [ memcached_out_of_cache_space_time ](https://github.com/netdata/netdata/blob/master/src/health/health.d/memcached.conf) | memcached.cache | estimated time the cache will run out of space if the system continues to add data at the same rate as the past hour |
## Setup
@@ -107,7 +107,7 @@ The configuration file name for this integration is `python.d/memcached.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/memcached/memcached.chart.py b/src/collectors/python.d.plugin/memcached/memcached.chart.py
index adb9560b7..adb9560b7 100644
--- a/collectors/python.d.plugin/memcached/memcached.chart.py
+++ b/src/collectors/python.d.plugin/memcached/memcached.chart.py
diff --git a/collectors/python.d.plugin/memcached/memcached.conf b/src/collectors/python.d.plugin/memcached/memcached.conf
index 3286b4623..3286b4623 100644
--- a/collectors/python.d.plugin/memcached/memcached.conf
+++ b/src/collectors/python.d.plugin/memcached/memcached.conf
diff --git a/collectors/python.d.plugin/monit/README.md b/src/collectors/python.d.plugin/monit/README.md
index ac69496f4..ac69496f4 120000
--- a/collectors/python.d.plugin/monit/README.md
+++ b/src/collectors/python.d.plugin/monit/README.md
diff --git a/collectors/python.d.plugin/monit/integrations/monit.md b/src/collectors/python.d.plugin/monit/integrations/monit.md
index 18219141d..dacd6e3a6 100644
--- a/collectors/python.d.plugin/monit/integrations/monit.md
+++ b/src/collectors/python.d.plugin/monit/integrations/monit.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/monit/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/monit/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/monit/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/monit/metadata.yaml"
sidebar_label: "Monit"
learn_status: "Published"
-learn_rel_path: "Data Collection/Synthetic Checks"
+learn_rel_path: "Collecting Metrics/Synthetic Checks"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -99,7 +99,7 @@ The configuration file name for this integration is `python.d/monit.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/monit/metadata.yaml b/src/collectors/python.d.plugin/monit/metadata.yaml
index b51273188..b51273188 100644
--- a/collectors/python.d.plugin/monit/metadata.yaml
+++ b/src/collectors/python.d.plugin/monit/metadata.yaml
diff --git a/collectors/python.d.plugin/monit/monit.chart.py b/src/collectors/python.d.plugin/monit/monit.chart.py
index 5d926961b..5d926961b 100644
--- a/collectors/python.d.plugin/monit/monit.chart.py
+++ b/src/collectors/python.d.plugin/monit/monit.chart.py
diff --git a/collectors/python.d.plugin/monit/monit.conf b/src/collectors/python.d.plugin/monit/monit.conf
index 9a3fb6938..9a3fb6938 100644
--- a/collectors/python.d.plugin/monit/monit.conf
+++ b/src/collectors/python.d.plugin/monit/monit.conf
diff --git a/collectors/python.d.plugin/nsd/README.md b/src/collectors/python.d.plugin/nsd/README.md
index 59fcfe491..59fcfe491 120000
--- a/collectors/python.d.plugin/nsd/README.md
+++ b/src/collectors/python.d.plugin/nsd/README.md
diff --git a/collectors/python.d.plugin/nsd/integrations/name_server_daemon.md b/src/collectors/python.d.plugin/nsd/integrations/name_server_daemon.md
index 0e66c44eb..e7ead13a1 100644
--- a/collectors/python.d.plugin/nsd/integrations/name_server_daemon.md
+++ b/src/collectors/python.d.plugin/nsd/integrations/name_server_daemon.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/nsd/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/nsd/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/nsd/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/nsd/metadata.yaml"
sidebar_label: "Name Server Daemon"
learn_status: "Published"
-learn_rel_path: "Data Collection/DNS and DHCP Servers"
+learn_rel_path: "Collecting Metrics/DNS and DHCP Servers"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -121,7 +121,7 @@ The configuration file name for this integration is `python.d/nsd.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/nsd/metadata.yaml b/src/collectors/python.d.plugin/nsd/metadata.yaml
index f5e2c46b0..f5e2c46b0 100644
--- a/collectors/python.d.plugin/nsd/metadata.yaml
+++ b/src/collectors/python.d.plugin/nsd/metadata.yaml
diff --git a/collectors/python.d.plugin/nsd/nsd.chart.py b/src/collectors/python.d.plugin/nsd/nsd.chart.py
index 6f9b2cec8..6f9b2cec8 100644
--- a/collectors/python.d.plugin/nsd/nsd.chart.py
+++ b/src/collectors/python.d.plugin/nsd/nsd.chart.py
diff --git a/collectors/python.d.plugin/nsd/nsd.conf b/src/collectors/python.d.plugin/nsd/nsd.conf
index 77a8a3177..77a8a3177 100644
--- a/collectors/python.d.plugin/nsd/nsd.conf
+++ b/src/collectors/python.d.plugin/nsd/nsd.conf
diff --git a/collectors/python.d.plugin/nvidia_smi/nvidia_smi.chart.py b/src/collectors/python.d.plugin/nvidia_smi/nvidia_smi.chart.py
index 556a61435..556a61435 100644
--- a/collectors/python.d.plugin/nvidia_smi/nvidia_smi.chart.py
+++ b/src/collectors/python.d.plugin/nvidia_smi/nvidia_smi.chart.py
diff --git a/collectors/python.d.plugin/nvidia_smi/nvidia_smi.conf b/src/collectors/python.d.plugin/nvidia_smi/nvidia_smi.conf
index 3d2a30d41..3d2a30d41 100644
--- a/collectors/python.d.plugin/nvidia_smi/nvidia_smi.conf
+++ b/src/collectors/python.d.plugin/nvidia_smi/nvidia_smi.conf
diff --git a/collectors/python.d.plugin/openldap/README.md b/src/collectors/python.d.plugin/openldap/README.md
index 45f36b9b9..45f36b9b9 120000
--- a/collectors/python.d.plugin/openldap/README.md
+++ b/src/collectors/python.d.plugin/openldap/README.md
diff --git a/collectors/python.d.plugin/openldap/integrations/openldap.md b/src/collectors/python.d.plugin/openldap/integrations/openldap.md
index a9480a490..b297df413 100644
--- a/collectors/python.d.plugin/openldap/integrations/openldap.md
+++ b/src/collectors/python.d.plugin/openldap/integrations/openldap.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/openldap/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/openldap/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/openldap/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/openldap/metadata.yaml"
sidebar_label: "OpenLDAP"
learn_status: "Published"
-learn_rel_path: "Data Collection/Authentication and Authorization"
+learn_rel_path: "Collecting Metrics/Authentication and Authorization"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -129,7 +129,7 @@ The configuration file name for this integration is `python.d/openldap.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/openldap/metadata.yaml b/src/collectors/python.d.plugin/openldap/metadata.yaml
index 3826b22c7..3826b22c7 100644
--- a/collectors/python.d.plugin/openldap/metadata.yaml
+++ b/src/collectors/python.d.plugin/openldap/metadata.yaml
diff --git a/collectors/python.d.plugin/openldap/openldap.chart.py b/src/collectors/python.d.plugin/openldap/openldap.chart.py
index aba143954..aba143954 100644
--- a/collectors/python.d.plugin/openldap/openldap.chart.py
+++ b/src/collectors/python.d.plugin/openldap/openldap.chart.py
diff --git a/collectors/python.d.plugin/openldap/openldap.conf b/src/collectors/python.d.plugin/openldap/openldap.conf
index 5fd99a525..5fd99a525 100644
--- a/collectors/python.d.plugin/openldap/openldap.conf
+++ b/src/collectors/python.d.plugin/openldap/openldap.conf
diff --git a/collectors/python.d.plugin/oracledb/README.md b/src/collectors/python.d.plugin/oracledb/README.md
index a75e3611e..a75e3611e 120000
--- a/collectors/python.d.plugin/oracledb/README.md
+++ b/src/collectors/python.d.plugin/oracledb/README.md
diff --git a/collectors/python.d.plugin/oracledb/integrations/oracle_db.md b/src/collectors/python.d.plugin/oracledb/integrations/oracle_db.md
index 30557c021..7290fe3f5 100644
--- a/collectors/python.d.plugin/oracledb/integrations/oracle_db.md
+++ b/src/collectors/python.d.plugin/oracledb/integrations/oracle_db.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/oracledb/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/oracledb/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/oracledb/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/oracledb/metadata.yaml"
sidebar_label: "Oracle DB"
learn_status: "Published"
-learn_rel_path: "Data Collection/Databases"
+learn_rel_path: "Collecting Metrics/Databases"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -137,7 +137,7 @@ The configuration file name for this integration is `python.d/oracledb.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/oracledb/metadata.yaml b/src/collectors/python.d.plugin/oracledb/metadata.yaml
index f2ab8312b..f2ab8312b 100644
--- a/collectors/python.d.plugin/oracledb/metadata.yaml
+++ b/src/collectors/python.d.plugin/oracledb/metadata.yaml
diff --git a/collectors/python.d.plugin/oracledb/oracledb.chart.py b/src/collectors/python.d.plugin/oracledb/oracledb.chart.py
index 455cf270e..455cf270e 100644
--- a/collectors/python.d.plugin/oracledb/oracledb.chart.py
+++ b/src/collectors/python.d.plugin/oracledb/oracledb.chart.py
diff --git a/collectors/python.d.plugin/oracledb/oracledb.conf b/src/collectors/python.d.plugin/oracledb/oracledb.conf
index 027215dad..027215dad 100644
--- a/collectors/python.d.plugin/oracledb/oracledb.conf
+++ b/src/collectors/python.d.plugin/oracledb/oracledb.conf
diff --git a/collectors/python.d.plugin/pandas/README.md b/src/collectors/python.d.plugin/pandas/README.md
index 2fabe63c1..2fabe63c1 120000
--- a/collectors/python.d.plugin/pandas/README.md
+++ b/src/collectors/python.d.plugin/pandas/README.md
diff --git a/collectors/python.d.plugin/pandas/integrations/pandas.md b/src/collectors/python.d.plugin/pandas/integrations/pandas.md
index 83c5c66b1..7fdae938e 100644
--- a/collectors/python.d.plugin/pandas/integrations/pandas.md
+++ b/src/collectors/python.d.plugin/pandas/integrations/pandas.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/pandas/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/pandas/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/pandas/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/pandas/metadata.yaml"
sidebar_label: "Pandas"
learn_status: "Published"
-learn_rel_path: "Data Collection/Generic Data Collection"
+learn_rel_path: "Collecting Metrics/Generic Collecting Metrics"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -109,7 +109,7 @@ The configuration file name for this integration is `python.d/pandas.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/pandas/metadata.yaml b/src/collectors/python.d.plugin/pandas/metadata.yaml
index 92ee1e986..92ee1e986 100644
--- a/collectors/python.d.plugin/pandas/metadata.yaml
+++ b/src/collectors/python.d.plugin/pandas/metadata.yaml
diff --git a/collectors/python.d.plugin/pandas/pandas.chart.py b/src/collectors/python.d.plugin/pandas/pandas.chart.py
index 7977bcb36..7977bcb36 100644
--- a/collectors/python.d.plugin/pandas/pandas.chart.py
+++ b/src/collectors/python.d.plugin/pandas/pandas.chart.py
diff --git a/collectors/python.d.plugin/pandas/pandas.conf b/src/collectors/python.d.plugin/pandas/pandas.conf
index 74a7da3e9..74a7da3e9 100644
--- a/collectors/python.d.plugin/pandas/pandas.conf
+++ b/src/collectors/python.d.plugin/pandas/pandas.conf
diff --git a/collectors/python.d.plugin/postfix/README.md b/src/collectors/python.d.plugin/postfix/README.md
index c62eb5c24..c62eb5c24 120000
--- a/collectors/python.d.plugin/postfix/README.md
+++ b/src/collectors/python.d.plugin/postfix/README.md
diff --git a/collectors/python.d.plugin/postfix/integrations/postfix.md b/src/collectors/python.d.plugin/postfix/integrations/postfix.md
index 2bb99922c..313fc9a31 100644
--- a/collectors/python.d.plugin/postfix/integrations/postfix.md
+++ b/src/collectors/python.d.plugin/postfix/integrations/postfix.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/postfix/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/postfix/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/postfix/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/postfix/metadata.yaml"
sidebar_label: "Postfix"
learn_status: "Published"
-learn_rel_path: "Data Collection/Mail Servers"
+learn_rel_path: "Collecting Metrics/Mail Servers"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
diff --git a/collectors/python.d.plugin/postfix/metadata.yaml b/src/collectors/python.d.plugin/postfix/metadata.yaml
index 1bbb61164..1bbb61164 100644
--- a/collectors/python.d.plugin/postfix/metadata.yaml
+++ b/src/collectors/python.d.plugin/postfix/metadata.yaml
diff --git a/collectors/python.d.plugin/postfix/postfix.chart.py b/src/collectors/python.d.plugin/postfix/postfix.chart.py
index b650514ee..b650514ee 100644
--- a/collectors/python.d.plugin/postfix/postfix.chart.py
+++ b/src/collectors/python.d.plugin/postfix/postfix.chart.py
diff --git a/collectors/python.d.plugin/postfix/postfix.conf b/src/collectors/python.d.plugin/postfix/postfix.conf
index a4d2472ee..a4d2472ee 100644
--- a/collectors/python.d.plugin/postfix/postfix.conf
+++ b/src/collectors/python.d.plugin/postfix/postfix.conf
diff --git a/collectors/python.d.plugin/puppet/README.md b/src/collectors/python.d.plugin/puppet/README.md
index b6c4c83f9..b6c4c83f9 120000
--- a/collectors/python.d.plugin/puppet/README.md
+++ b/src/collectors/python.d.plugin/puppet/README.md
diff --git a/collectors/python.d.plugin/puppet/integrations/puppet.md b/src/collectors/python.d.plugin/puppet/integrations/puppet.md
index ca190b576..31cff9888 100644
--- a/collectors/python.d.plugin/puppet/integrations/puppet.md
+++ b/src/collectors/python.d.plugin/puppet/integrations/puppet.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/puppet/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/puppet/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/puppet/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/puppet/metadata.yaml"
sidebar_label: "Puppet"
learn_status: "Published"
-learn_rel_path: "Data Collection/CICD Platforms"
+learn_rel_path: "Collecting Metrics/CICD Platforms"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -91,7 +91,7 @@ The configuration file name for this integration is `python.d/puppet.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/puppet/metadata.yaml b/src/collectors/python.d.plugin/puppet/metadata.yaml
index 781519b6a..781519b6a 100644
--- a/collectors/python.d.plugin/puppet/metadata.yaml
+++ b/src/collectors/python.d.plugin/puppet/metadata.yaml
diff --git a/collectors/python.d.plugin/puppet/puppet.chart.py b/src/collectors/python.d.plugin/puppet/puppet.chart.py
index f8adf6006..f8adf6006 100644
--- a/collectors/python.d.plugin/puppet/puppet.chart.py
+++ b/src/collectors/python.d.plugin/puppet/puppet.chart.py
diff --git a/collectors/python.d.plugin/puppet/puppet.conf b/src/collectors/python.d.plugin/puppet/puppet.conf
index ff5c3d020..ff5c3d020 100644
--- a/collectors/python.d.plugin/puppet/puppet.conf
+++ b/src/collectors/python.d.plugin/puppet/puppet.conf
diff --git a/collectors/python.d.plugin/python.d.conf b/src/collectors/python.d.plugin/python.d.conf
index 3953ce2b4..3953ce2b4 100644
--- a/collectors/python.d.plugin/python.d.conf
+++ b/src/collectors/python.d.plugin/python.d.conf
diff --git a/collectors/python.d.plugin/python.d.plugin.in b/src/collectors/python.d.plugin/python.d.plugin.in
index 86fea209c..86fea209c 100644
--- a/collectors/python.d.plugin/python.d.plugin.in
+++ b/src/collectors/python.d.plugin/python.d.plugin.in
diff --git a/collectors/python.d.plugin/python_modules/__init__.py b/src/collectors/python.d.plugin/python_modules/__init__.py
index e69de29bb..e69de29bb 100644
--- a/collectors/python.d.plugin/python_modules/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/ExecutableService.py b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/ExecutableService.py
index a74b4239e..a74b4239e 100644
--- a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/ExecutableService.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/ExecutableService.py
diff --git a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/LogService.py b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/LogService.py
index a55e33f52..a55e33f52 100644
--- a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/LogService.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/LogService.py
diff --git a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/MySQLService.py b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/MySQLService.py
index 7f5c7d221..7f5c7d221 100644
--- a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/MySQLService.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/MySQLService.py
diff --git a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/SimpleService.py b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/SimpleService.py
index 3f122e1d9..3f122e1d9 100644
--- a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/SimpleService.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/SimpleService.py
diff --git a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/SocketService.py b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/SocketService.py
index d6c755058..d6c755058 100644
--- a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/SocketService.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/SocketService.py
diff --git a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/UrlService.py b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/UrlService.py
index 76129d376..76129d376 100644
--- a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/UrlService.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/UrlService.py
diff --git a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/__init__.py b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/__init__.py
index e69de29bb..e69de29bb 100644
--- a/collectors/python.d.plugin/python_modules/bases/FrameworkServices/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/FrameworkServices/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/bases/__init__.py b/src/collectors/python.d.plugin/python_modules/bases/__init__.py
index e69de29bb..e69de29bb 100644
--- a/collectors/python.d.plugin/python_modules/bases/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/bases/charts.py b/src/collectors/python.d.plugin/python_modules/bases/charts.py
index 203ad1672..203ad1672 100644
--- a/collectors/python.d.plugin/python_modules/bases/charts.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/charts.py
diff --git a/collectors/python.d.plugin/python_modules/bases/collection.py b/src/collectors/python.d.plugin/python_modules/bases/collection.py
index 93bf8cf05..93bf8cf05 100644
--- a/collectors/python.d.plugin/python_modules/bases/collection.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/collection.py
diff --git a/collectors/python.d.plugin/python_modules/bases/loaders.py b/src/collectors/python.d.plugin/python_modules/bases/loaders.py
index 095f3a3b1..095f3a3b1 100644
--- a/collectors/python.d.plugin/python_modules/bases/loaders.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/loaders.py
diff --git a/collectors/python.d.plugin/python_modules/bases/loggers.py b/src/collectors/python.d.plugin/python_modules/bases/loggers.py
index 7ae8ab0c1..7ae8ab0c1 100644
--- a/collectors/python.d.plugin/python_modules/bases/loggers.py
+++ b/src/collectors/python.d.plugin/python_modules/bases/loggers.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/__init__.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/__init__.py
index 4d560e438..4d560e438 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/composer.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/composer.py
index 6b41b8067..6b41b8067 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/composer.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/composer.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/constructor.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/constructor.py
index 8ad1b90a7..8ad1b90a7 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/constructor.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/constructor.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/cyaml.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/cyaml.py
index 2858ab479..2858ab479 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/cyaml.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/cyaml.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/dumper.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/dumper.py
index 3685cbeeb..3685cbeeb 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/dumper.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/dumper.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/emitter.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/emitter.py
index 9a460a0fd..9a460a0fd 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/emitter.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/emitter.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/error.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/error.py
index 5466be721..5466be721 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/error.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/error.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/events.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/events.py
index 283452add..283452add 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/events.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/events.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/loader.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/loader.py
index 1c195531f..1c195531f 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/loader.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/loader.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/nodes.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/nodes.py
index ed2a1b43e..ed2a1b43e 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/nodes.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/nodes.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/parser.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/parser.py
index 97ba08337..97ba08337 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/parser.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/parser.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/reader.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/reader.py
index 8d422954e..8d422954e 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/reader.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/reader.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/representer.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/representer.py
index 0a1404eca..0a1404eca 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/representer.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/representer.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/resolver.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/resolver.py
index 49922debf..49922debf 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/resolver.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/resolver.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/scanner.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/scanner.py
index 971da6127..971da6127 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/scanner.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/scanner.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/serializer.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/serializer.py
index 15fdbb0c0..15fdbb0c0 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/serializer.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/serializer.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml2/tokens.py b/src/collectors/python.d.plugin/python_modules/pyyaml2/tokens.py
index c5c4fb116..c5c4fb116 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml2/tokens.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml2/tokens.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/__init__.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/__init__.py
index a884b33cf..a884b33cf 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/composer.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/composer.py
index c418bba91..c418bba91 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/composer.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/composer.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/constructor.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/constructor.py
index ee09a7a7e..ee09a7a7e 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/constructor.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/constructor.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/cyaml.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/cyaml.py
index e6c16d894..e6c16d894 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/cyaml.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/cyaml.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/dumper.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/dumper.py
index ba590c6e6..ba590c6e6 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/dumper.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/dumper.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/emitter.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/emitter.py
index d4be65a8e..d4be65a8e 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/emitter.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/emitter.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/error.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/error.py
index 5fec7d449..5fec7d449 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/error.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/error.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/events.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/events.py
index 283452add..283452add 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/events.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/events.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/loader.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/loader.py
index 7ef6cf815..7ef6cf815 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/loader.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/loader.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/nodes.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/nodes.py
index ed2a1b43e..ed2a1b43e 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/nodes.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/nodes.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/parser.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/parser.py
index bcec7f994..bcec7f994 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/parser.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/parser.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/reader.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/reader.py
index 0a515fd64..0a515fd64 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/reader.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/reader.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/representer.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/representer.py
index 756a18dcc..756a18dcc 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/representer.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/representer.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/resolver.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/resolver.py
index 50945e04d..50945e04d 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/resolver.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/resolver.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/scanner.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/scanner.py
index b55854e8b..b55854e8b 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/scanner.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/scanner.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/serializer.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/serializer.py
index 1ba2f7f9d..1ba2f7f9d 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/serializer.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/serializer.py
diff --git a/collectors/python.d.plugin/python_modules/pyyaml3/tokens.py b/src/collectors/python.d.plugin/python_modules/pyyaml3/tokens.py
index c5c4fb116..c5c4fb116 100644
--- a/collectors/python.d.plugin/python_modules/pyyaml3/tokens.py
+++ b/src/collectors/python.d.plugin/python_modules/pyyaml3/tokens.py
diff --git a/collectors/python.d.plugin/python_modules/third_party/__init__.py b/src/collectors/python.d.plugin/python_modules/third_party/__init__.py
index e69de29bb..e69de29bb 100644
--- a/collectors/python.d.plugin/python_modules/third_party/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/third_party/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/third_party/boinc_client.py b/src/collectors/python.d.plugin/python_modules/third_party/boinc_client.py
index ec21779a0..ec21779a0 100644
--- a/collectors/python.d.plugin/python_modules/third_party/boinc_client.py
+++ b/src/collectors/python.d.plugin/python_modules/third_party/boinc_client.py
diff --git a/collectors/python.d.plugin/python_modules/third_party/filelock.py b/src/collectors/python.d.plugin/python_modules/third_party/filelock.py
index 4c981672b..4c981672b 100644
--- a/collectors/python.d.plugin/python_modules/third_party/filelock.py
+++ b/src/collectors/python.d.plugin/python_modules/third_party/filelock.py
diff --git a/collectors/python.d.plugin/python_modules/third_party/lm_sensors.py b/src/collectors/python.d.plugin/python_modules/third_party/lm_sensors.py
index f873eac83..f873eac83 100644
--- a/collectors/python.d.plugin/python_modules/third_party/lm_sensors.py
+++ b/src/collectors/python.d.plugin/python_modules/third_party/lm_sensors.py
diff --git a/collectors/python.d.plugin/python_modules/third_party/mcrcon.py b/src/collectors/python.d.plugin/python_modules/third_party/mcrcon.py
index a65a304b6..a65a304b6 100644
--- a/collectors/python.d.plugin/python_modules/third_party/mcrcon.py
+++ b/src/collectors/python.d.plugin/python_modules/third_party/mcrcon.py
diff --git a/collectors/python.d.plugin/python_modules/third_party/monotonic.py b/src/collectors/python.d.plugin/python_modules/third_party/monotonic.py
index 4ebd556c3..4ebd556c3 100644
--- a/collectors/python.d.plugin/python_modules/third_party/monotonic.py
+++ b/src/collectors/python.d.plugin/python_modules/third_party/monotonic.py
diff --git a/collectors/python.d.plugin/python_modules/third_party/ordereddict.py b/src/collectors/python.d.plugin/python_modules/third_party/ordereddict.py
index 589401b8f..589401b8f 100644
--- a/collectors/python.d.plugin/python_modules/third_party/ordereddict.py
+++ b/src/collectors/python.d.plugin/python_modules/third_party/ordereddict.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/__init__.py b/src/collectors/python.d.plugin/python_modules/urllib3/__init__.py
index 3add84816..3add84816 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/_collections.py b/src/collectors/python.d.plugin/python_modules/urllib3/_collections.py
index 2a6b3ec70..2a6b3ec70 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/_collections.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/_collections.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/connection.py b/src/collectors/python.d.plugin/python_modules/urllib3/connection.py
index f757493c7..f757493c7 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/connection.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/connection.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/connectionpool.py b/src/collectors/python.d.plugin/python_modules/urllib3/connectionpool.py
index 90e4c86a5..90e4c86a5 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/connectionpool.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/connectionpool.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/contrib/__init__.py b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/__init__.py
index e69de29bb..e69de29bb 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/contrib/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/__init__.py b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/__init__.py
index e69de29bb..e69de29bb 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/bindings.py b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/bindings.py
index bb826673f..bb826673f 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/bindings.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/bindings.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/low_level.py b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/low_level.py
index 0f79a1372..0f79a1372 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/low_level.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/_securetransport/low_level.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/contrib/appengine.py b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/appengine.py
index e74589fa8..e74589fa8 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/contrib/appengine.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/appengine.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/contrib/ntlmpool.py b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/ntlmpool.py
index 3f8c9ebf5..3f8c9ebf5 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/contrib/ntlmpool.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/ntlmpool.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/contrib/pyopenssl.py b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/pyopenssl.py
index 8d373507d..8d373507d 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/contrib/pyopenssl.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/pyopenssl.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/contrib/securetransport.py b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/securetransport.py
index fcc30118c..fcc30118c 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/contrib/securetransport.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/securetransport.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/contrib/socks.py b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/socks.py
index 1cb79285b..1cb79285b 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/contrib/socks.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/contrib/socks.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/exceptions.py b/src/collectors/python.d.plugin/python_modules/urllib3/exceptions.py
index a71cabe06..a71cabe06 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/exceptions.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/exceptions.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/fields.py b/src/collectors/python.d.plugin/python_modules/urllib3/fields.py
index de7577b74..de7577b74 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/fields.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/fields.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/filepost.py b/src/collectors/python.d.plugin/python_modules/urllib3/filepost.py
index 3febc9cfe..3febc9cfe 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/filepost.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/filepost.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/packages/__init__.py b/src/collectors/python.d.plugin/python_modules/urllib3/packages/__init__.py
index 170e974c1..170e974c1 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/packages/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/packages/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/packages/backports/__init__.py b/src/collectors/python.d.plugin/python_modules/urllib3/packages/backports/__init__.py
index e69de29bb..e69de29bb 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/packages/backports/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/packages/backports/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/packages/backports/makefile.py b/src/collectors/python.d.plugin/python_modules/urllib3/packages/backports/makefile.py
index 8ab122f8b..8ab122f8b 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/packages/backports/makefile.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/packages/backports/makefile.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/packages/ordered_dict.py b/src/collectors/python.d.plugin/python_modules/urllib3/packages/ordered_dict.py
index 9f7c0e6b8..9f7c0e6b8 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/packages/ordered_dict.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/packages/ordered_dict.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/packages/six.py b/src/collectors/python.d.plugin/python_modules/urllib3/packages/six.py
index 31df5012b..31df5012b 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/packages/six.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/packages/six.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/__init__.py b/src/collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/__init__.py
index 2aeeeff91..2aeeeff91 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/_implementation.py b/src/collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/_implementation.py
index 647e081da..647e081da 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/_implementation.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/packages/ssl_match_hostname/_implementation.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/poolmanager.py b/src/collectors/python.d.plugin/python_modules/urllib3/poolmanager.py
index adea9bc01..adea9bc01 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/poolmanager.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/poolmanager.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/request.py b/src/collectors/python.d.plugin/python_modules/urllib3/request.py
index f78331975..f78331975 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/request.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/request.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/response.py b/src/collectors/python.d.plugin/python_modules/urllib3/response.py
index cf14a3076..cf14a3076 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/response.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/response.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/util/__init__.py b/src/collectors/python.d.plugin/python_modules/urllib3/util/__init__.py
index bba628d98..bba628d98 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/util/__init__.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/util/__init__.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/util/connection.py b/src/collectors/python.d.plugin/python_modules/urllib3/util/connection.py
index 3bd69e8fa..3bd69e8fa 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/util/connection.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/util/connection.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/util/request.py b/src/collectors/python.d.plugin/python_modules/urllib3/util/request.py
index 18f27b032..18f27b032 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/util/request.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/util/request.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/util/response.py b/src/collectors/python.d.plugin/python_modules/urllib3/util/response.py
index e4cda93d4..e4cda93d4 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/util/response.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/util/response.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/util/retry.py b/src/collectors/python.d.plugin/python_modules/urllib3/util/retry.py
index 61e63afec..61e63afec 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/util/retry.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/util/retry.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/util/selectors.py b/src/collectors/python.d.plugin/python_modules/urllib3/util/selectors.py
index de5e49838..de5e49838 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/util/selectors.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/util/selectors.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/util/ssl_.py b/src/collectors/python.d.plugin/python_modules/urllib3/util/ssl_.py
index ece3ec39e..ece3ec39e 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/util/ssl_.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/util/ssl_.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/util/timeout.py b/src/collectors/python.d.plugin/python_modules/urllib3/util/timeout.py
index 4041cf9b9..4041cf9b9 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/util/timeout.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/util/timeout.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/util/url.py b/src/collectors/python.d.plugin/python_modules/urllib3/util/url.py
index 99fd6534a..99fd6534a 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/util/url.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/util/url.py
diff --git a/collectors/python.d.plugin/python_modules/urllib3/util/wait.py b/src/collectors/python.d.plugin/python_modules/urllib3/util/wait.py
index 21e72979c..21e72979c 100644
--- a/collectors/python.d.plugin/python_modules/urllib3/util/wait.py
+++ b/src/collectors/python.d.plugin/python_modules/urllib3/util/wait.py
diff --git a/collectors/python.d.plugin/rethinkdbs/README.md b/src/collectors/python.d.plugin/rethinkdbs/README.md
index 78ddcfa18..78ddcfa18 120000
--- a/collectors/python.d.plugin/rethinkdbs/README.md
+++ b/src/collectors/python.d.plugin/rethinkdbs/README.md
diff --git a/collectors/python.d.plugin/rethinkdbs/integrations/rethinkdb.md b/src/collectors/python.d.plugin/rethinkdbs/integrations/rethinkdb.md
index ab51c0514..d7120c403 100644
--- a/collectors/python.d.plugin/rethinkdbs/integrations/rethinkdb.md
+++ b/src/collectors/python.d.plugin/rethinkdbs/integrations/rethinkdb.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/rethinkdbs/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/rethinkdbs/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/rethinkdbs/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/rethinkdbs/metadata.yaml"
sidebar_label: "RethinkDB"
learn_status: "Published"
-learn_rel_path: "Data Collection/Databases"
+learn_rel_path: "Collecting Metrics/Databases"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -107,7 +107,7 @@ The configuration file name for this integration is `python.d/rethinkdbs.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/rethinkdbs/metadata.yaml b/src/collectors/python.d.plugin/rethinkdbs/metadata.yaml
index bbc50eac6..bbc50eac6 100644
--- a/collectors/python.d.plugin/rethinkdbs/metadata.yaml
+++ b/src/collectors/python.d.plugin/rethinkdbs/metadata.yaml
diff --git a/collectors/python.d.plugin/rethinkdbs/rethinkdbs.chart.py b/src/collectors/python.d.plugin/rethinkdbs/rethinkdbs.chart.py
index e3fbc3632..e3fbc3632 100644
--- a/collectors/python.d.plugin/rethinkdbs/rethinkdbs.chart.py
+++ b/src/collectors/python.d.plugin/rethinkdbs/rethinkdbs.chart.py
diff --git a/collectors/python.d.plugin/rethinkdbs/rethinkdbs.conf b/src/collectors/python.d.plugin/rethinkdbs/rethinkdbs.conf
index d671acbb0..d671acbb0 100644
--- a/collectors/python.d.plugin/rethinkdbs/rethinkdbs.conf
+++ b/src/collectors/python.d.plugin/rethinkdbs/rethinkdbs.conf
diff --git a/collectors/python.d.plugin/retroshare/README.md b/src/collectors/python.d.plugin/retroshare/README.md
index 4e4c2cdb7..4e4c2cdb7 120000
--- a/collectors/python.d.plugin/retroshare/README.md
+++ b/src/collectors/python.d.plugin/retroshare/README.md
diff --git a/collectors/python.d.plugin/retroshare/integrations/retroshare.md b/src/collectors/python.d.plugin/retroshare/integrations/retroshare.md
index 4fc003c6f..12b0aa131 100644
--- a/collectors/python.d.plugin/retroshare/integrations/retroshare.md
+++ b/src/collectors/python.d.plugin/retroshare/integrations/retroshare.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/retroshare/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/retroshare/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/retroshare/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/retroshare/metadata.yaml"
sidebar_label: "RetroShare"
learn_status: "Published"
-learn_rel_path: "Data Collection/Media Services"
+learn_rel_path: "Collecting Metrics/Media Services"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -76,7 +76,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ retroshare_dht_working ](https://github.com/netdata/netdata/blob/master/health/health.d/retroshare.conf) | retroshare.dht | number of DHT peers |
+| [ retroshare_dht_working ](https://github.com/netdata/netdata/blob/master/src/health/health.d/retroshare.conf) | retroshare.dht | number of DHT peers |
## Setup
@@ -97,7 +97,7 @@ The configuration file name for this integration is `python.d/retroshare.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/retroshare/retroshare.chart.py b/src/collectors/python.d.plugin/retroshare/retroshare.chart.py
index 3f9593e94..3f9593e94 100644
--- a/collectors/python.d.plugin/retroshare/retroshare.chart.py
+++ b/src/collectors/python.d.plugin/retroshare/retroshare.chart.py
diff --git a/collectors/python.d.plugin/retroshare/retroshare.conf b/src/collectors/python.d.plugin/retroshare/retroshare.conf
index 3d0af538d..3d0af538d 100644
--- a/collectors/python.d.plugin/retroshare/retroshare.conf
+++ b/src/collectors/python.d.plugin/retroshare/retroshare.conf
diff --git a/collectors/python.d.plugin/riakkv/README.md b/src/collectors/python.d.plugin/riakkv/README.md
index f43ece09b..f43ece09b 120000
--- a/collectors/python.d.plugin/riakkv/README.md
+++ b/src/collectors/python.d.plugin/riakkv/README.md
diff --git a/collectors/python.d.plugin/riakkv/integrations/riakkv.md b/src/collectors/python.d.plugin/riakkv/integrations/riakkv.md
index 2e8279bc3..ef4ea080b 100644
--- a/collectors/python.d.plugin/riakkv/integrations/riakkv.md
+++ b/src/collectors/python.d.plugin/riakkv/integrations/riakkv.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/riakkv/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/riakkv/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/riakkv/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/riakkv/metadata.yaml"
sidebar_label: "RiakKV"
learn_status: "Published"
-learn_rel_path: "Data Collection/Databases"
+learn_rel_path: "Collecting Metrics/Databases"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -99,12 +99,12 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ riakkv_1h_kv_get_mean_latency ](https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf) | riak.kv.latency.get | average time between reception of client GET request and subsequent response to client over the last hour |
-| [ riakkv_kv_get_slow ](https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf) | riak.kv.latency.get | average time between reception of client GET request and subsequent response to the client over the last 3 minutes, compared to the average over the last hour |
-| [ riakkv_1h_kv_put_mean_latency ](https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf) | riak.kv.latency.put | average time between reception of client PUT request and subsequent response to the client over the last hour |
-| [ riakkv_kv_put_slow ](https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf) | riak.kv.latency.put | average time between reception of client PUT request and subsequent response to the client over the last 3 minutes, compared to the average over the last hour |
-| [ riakkv_vm_high_process_count ](https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf) | riak.vm | number of processes running in the Erlang VM |
-| [ riakkv_list_keys_active ](https://github.com/netdata/netdata/blob/master/health/health.d/riakkv.conf) | riak.core.fsm_active | number of currently running list keys finite state machines |
+| [ riakkv_1h_kv_get_mean_latency ](https://github.com/netdata/netdata/blob/master/src/health/health.d/riakkv.conf) | riak.kv.latency.get | average time between reception of client GET request and subsequent response to client over the last hour |
+| [ riakkv_kv_get_slow ](https://github.com/netdata/netdata/blob/master/src/health/health.d/riakkv.conf) | riak.kv.latency.get | average time between reception of client GET request and subsequent response to the client over the last 3 minutes, compared to the average over the last hour |
+| [ riakkv_1h_kv_put_mean_latency ](https://github.com/netdata/netdata/blob/master/src/health/health.d/riakkv.conf) | riak.kv.latency.put | average time between reception of client PUT request and subsequent response to the client over the last hour |
+| [ riakkv_kv_put_slow ](https://github.com/netdata/netdata/blob/master/src/health/health.d/riakkv.conf) | riak.kv.latency.put | average time between reception of client PUT request and subsequent response to the client over the last 3 minutes, compared to the average over the last hour |
+| [ riakkv_vm_high_process_count ](https://github.com/netdata/netdata/blob/master/src/health/health.d/riakkv.conf) | riak.vm | number of processes running in the Erlang VM |
+| [ riakkv_list_keys_active ](https://github.com/netdata/netdata/blob/master/src/health/health.d/riakkv.conf) | riak.core.fsm_active | number of currently running list keys finite state machines |
## Setup
@@ -127,7 +127,7 @@ The configuration file name for this integration is `python.d/riakkv.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/riakkv/riakkv.chart.py b/src/collectors/python.d.plugin/riakkv/riakkv.chart.py
index c390c8bc0..c390c8bc0 100644
--- a/collectors/python.d.plugin/riakkv/riakkv.chart.py
+++ b/src/collectors/python.d.plugin/riakkv/riakkv.chart.py
diff --git a/collectors/python.d.plugin/riakkv/riakkv.conf b/src/collectors/python.d.plugin/riakkv/riakkv.conf
index be01c48ac..be01c48ac 100644
--- a/collectors/python.d.plugin/riakkv/riakkv.conf
+++ b/src/collectors/python.d.plugin/riakkv/riakkv.conf
diff --git a/collectors/python.d.plugin/samba/README.md b/src/collectors/python.d.plugin/samba/README.md
index 3b63bbab6..3b63bbab6 120000
--- a/collectors/python.d.plugin/samba/README.md
+++ b/src/collectors/python.d.plugin/samba/README.md
diff --git a/collectors/python.d.plugin/samba/integrations/samba.md b/src/collectors/python.d.plugin/samba/integrations/samba.md
index 1bd1664ee..e34dd03e3 100644
--- a/collectors/python.d.plugin/samba/integrations/samba.md
+++ b/src/collectors/python.d.plugin/samba/integrations/samba.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/samba/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/samba/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/samba/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/samba/metadata.yaml"
sidebar_label: "Samba"
learn_status: "Published"
-learn_rel_path: "Data Collection/Storage, Mount Points and Filesystems"
+learn_rel_path: "Collecting Metrics/Storage, Mount Points and Filesystems"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -91,13 +91,13 @@ There are no alerts configured by default for this integration.
#### Enable the samba collector
-The `samba` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md), which is typically at `/etc/netdata`, to edit the `python.d.conf` file.
+The `samba` collector is disabled by default. To enable it, use `edit-config` from the Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md), which is typically at `/etc/netdata`, to edit the `python.d.conf` file.
```bash
cd /etc/netdata # Replace this path with your Netdata config directory, if different
sudo ./edit-config python.d.conf
```
-Change the value of the `samba` setting to `yes`. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md) for your system.
+Change the value of the `samba` setting to `yes`. Save the file and restart the Netdata Agent with `sudo systemctl restart netdata`, or the [appropriate method](https://github.com/netdata/netdata/blob/master/packaging/installer/README.md#maintaining-a-netdata-agent-installation) for your system.
#### Permissions and programs
@@ -143,7 +143,7 @@ The configuration file name for this integration is `python.d/samba.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/samba/samba.chart.py b/src/collectors/python.d.plugin/samba/samba.chart.py
index 8eebcd60c..8eebcd60c 100644
--- a/collectors/python.d.plugin/samba/samba.chart.py
+++ b/src/collectors/python.d.plugin/samba/samba.chart.py
diff --git a/collectors/python.d.plugin/samba/samba.conf b/src/collectors/python.d.plugin/samba/samba.conf
index db15d4e9e..db15d4e9e 100644
--- a/collectors/python.d.plugin/samba/samba.conf
+++ b/src/collectors/python.d.plugin/samba/samba.conf
diff --git a/collectors/python.d.plugin/sensors/README.md b/src/collectors/python.d.plugin/sensors/README.md
index 4e92b0882..4e92b0882 120000
--- a/collectors/python.d.plugin/sensors/README.md
+++ b/src/collectors/python.d.plugin/sensors/README.md
diff --git a/collectors/python.d.plugin/sensors/integrations/linux_sensors_lm-sensors.md b/src/collectors/python.d.plugin/sensors/integrations/linux_sensors_lm-sensors.md
index e426c8c83..f743d410c 100644
--- a/collectors/python.d.plugin/sensors/integrations/linux_sensors_lm-sensors.md
+++ b/src/collectors/python.d.plugin/sensors/integrations/linux_sensors_lm-sensors.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/sensors/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/sensors/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/sensors/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/sensors/metadata.yaml"
sidebar_label: "Linux Sensors (lm-sensors)"
learn_status: "Published"
-learn_rel_path: "Data Collection/Hardware Devices and Sensors"
+learn_rel_path: "Collecting Metrics/Hardware Devices and Sensors"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -99,7 +99,7 @@ The configuration file name for this integration is `python.d/sensors.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/sensors/sensors.chart.py b/src/collectors/python.d.plugin/sensors/sensors.chart.py
index 0d9de3750..0d9de3750 100644
--- a/collectors/python.d.plugin/sensors/sensors.chart.py
+++ b/src/collectors/python.d.plugin/sensors/sensors.chart.py
diff --git a/collectors/python.d.plugin/sensors/sensors.conf b/src/collectors/python.d.plugin/sensors/sensors.conf
index d3369ba66..d3369ba66 100644
--- a/collectors/python.d.plugin/sensors/sensors.conf
+++ b/src/collectors/python.d.plugin/sensors/sensors.conf
diff --git a/collectors/python.d.plugin/smartd_log/README.md b/src/collectors/python.d.plugin/smartd_log/README.md
index 63aad6c85..63aad6c85 120000
--- a/collectors/python.d.plugin/smartd_log/README.md
+++ b/src/collectors/python.d.plugin/smartd_log/README.md
diff --git a/collectors/python.d.plugin/smartd_log/integrations/s.m.a.r.t..md b/src/collectors/python.d.plugin/smartd_log/integrations/s.m.a.r.t..md
index 5c5b569e9..73a96ae55 100644
--- a/collectors/python.d.plugin/smartd_log/integrations/s.m.a.r.t..md
+++ b/src/collectors/python.d.plugin/smartd_log/integrations/s.m.a.r.t..md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/smartd_log/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/smartd_log/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/smartd_log/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/smartd_log/metadata.yaml"
sidebar_label: "S.M.A.R.T."
learn_status: "Published"
-learn_rel_path: "Data Collection/Hardware Devices and Sensors"
+learn_rel_path: "Collecting Metrics/Hardware Devices and Sensors"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -143,7 +143,7 @@ The configuration file name for this integration is `python.d/smartd_log.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/smartd_log/metadata.yaml b/src/collectors/python.d.plugin/smartd_log/metadata.yaml
index d11949691..d11949691 100644
--- a/collectors/python.d.plugin/smartd_log/metadata.yaml
+++ b/src/collectors/python.d.plugin/smartd_log/metadata.yaml
diff --git a/collectors/python.d.plugin/smartd_log/smartd_log.chart.py b/src/collectors/python.d.plugin/smartd_log/smartd_log.chart.py
index a896164df..a896164df 100644
--- a/collectors/python.d.plugin/smartd_log/smartd_log.chart.py
+++ b/src/collectors/python.d.plugin/smartd_log/smartd_log.chart.py
diff --git a/collectors/python.d.plugin/smartd_log/smartd_log.conf b/src/collectors/python.d.plugin/smartd_log/smartd_log.conf
index 3e81317f1..3e81317f1 100644
--- a/collectors/python.d.plugin/smartd_log/smartd_log.conf
+++ b/src/collectors/python.d.plugin/smartd_log/smartd_log.conf
diff --git a/collectors/python.d.plugin/spigotmc/README.md b/src/collectors/python.d.plugin/spigotmc/README.md
index 66e5c9c47..66e5c9c47 120000
--- a/collectors/python.d.plugin/spigotmc/README.md
+++ b/src/collectors/python.d.plugin/spigotmc/README.md
diff --git a/collectors/python.d.plugin/spigotmc/integrations/spigotmc.md b/src/collectors/python.d.plugin/spigotmc/integrations/spigotmc.md
index 55ec8fa22..8880acca4 100644
--- a/collectors/python.d.plugin/spigotmc/integrations/spigotmc.md
+++ b/src/collectors/python.d.plugin/spigotmc/integrations/spigotmc.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/spigotmc/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/spigotmc/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/spigotmc/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/spigotmc/metadata.yaml"
sidebar_label: "SpigotMC"
learn_status: "Published"
-learn_rel_path: "Data Collection/Gaming"
+learn_rel_path: "Collecting Metrics/Gaming"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -98,7 +98,7 @@ The configuration file name for this integration is `python.d/spigotmc.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/spigotmc/metadata.yaml b/src/collectors/python.d.plugin/spigotmc/metadata.yaml
index 5dea9f0c8..5dea9f0c8 100644
--- a/collectors/python.d.plugin/spigotmc/metadata.yaml
+++ b/src/collectors/python.d.plugin/spigotmc/metadata.yaml
diff --git a/collectors/python.d.plugin/spigotmc/spigotmc.chart.py b/src/collectors/python.d.plugin/spigotmc/spigotmc.chart.py
index 81370fb4c..81370fb4c 100644
--- a/collectors/python.d.plugin/spigotmc/spigotmc.chart.py
+++ b/src/collectors/python.d.plugin/spigotmc/spigotmc.chart.py
diff --git a/collectors/python.d.plugin/spigotmc/spigotmc.conf b/src/collectors/python.d.plugin/spigotmc/spigotmc.conf
index f0064ea2f..f0064ea2f 100644
--- a/collectors/python.d.plugin/spigotmc/spigotmc.conf
+++ b/src/collectors/python.d.plugin/spigotmc/spigotmc.conf
diff --git a/collectors/python.d.plugin/squid/README.md b/src/collectors/python.d.plugin/squid/README.md
index c4e5a03d7..c4e5a03d7 120000
--- a/collectors/python.d.plugin/squid/README.md
+++ b/src/collectors/python.d.plugin/squid/README.md
diff --git a/collectors/python.d.plugin/squid/integrations/squid.md b/src/collectors/python.d.plugin/squid/integrations/squid.md
index 6599826da..d014898f6 100644
--- a/collectors/python.d.plugin/squid/integrations/squid.md
+++ b/src/collectors/python.d.plugin/squid/integrations/squid.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/squid/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/squid/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/squid/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/squid/metadata.yaml"
sidebar_label: "Squid"
learn_status: "Published"
-learn_rel_path: "Data Collection/Web Servers and Web Proxies"
+learn_rel_path: "Collecting Metrics/Web Servers and Web Proxies"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -94,7 +94,7 @@ The configuration file name for this integration is `python.d/squid.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/squid/metadata.yaml b/src/collectors/python.d.plugin/squid/metadata.yaml
index d0c5b3ecc..d0c5b3ecc 100644
--- a/collectors/python.d.plugin/squid/metadata.yaml
+++ b/src/collectors/python.d.plugin/squid/metadata.yaml
diff --git a/collectors/python.d.plugin/squid/squid.chart.py b/src/collectors/python.d.plugin/squid/squid.chart.py
index bcae2d892..bcae2d892 100644
--- a/collectors/python.d.plugin/squid/squid.chart.py
+++ b/src/collectors/python.d.plugin/squid/squid.chart.py
diff --git a/collectors/python.d.plugin/squid/squid.conf b/src/collectors/python.d.plugin/squid/squid.conf
index b90a52c0c..b90a52c0c 100644
--- a/collectors/python.d.plugin/squid/squid.conf
+++ b/src/collectors/python.d.plugin/squid/squid.conf
diff --git a/collectors/python.d.plugin/tomcat/README.md b/src/collectors/python.d.plugin/tomcat/README.md
index 997090c35..997090c35 120000
--- a/collectors/python.d.plugin/tomcat/README.md
+++ b/src/collectors/python.d.plugin/tomcat/README.md
diff --git a/collectors/python.d.plugin/tomcat/integrations/tomcat.md b/src/collectors/python.d.plugin/tomcat/integrations/tomcat.md
index 883f29dd3..9bfe8980e 100644
--- a/collectors/python.d.plugin/tomcat/integrations/tomcat.md
+++ b/src/collectors/python.d.plugin/tomcat/integrations/tomcat.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/tomcat/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/tomcat/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/tomcat/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/tomcat/metadata.yaml"
sidebar_label: "Tomcat"
learn_status: "Published"
-learn_rel_path: "Data Collection/Web Servers and Web Proxies"
+learn_rel_path: "Collecting Metrics/Web Servers and Web Proxies"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -99,7 +99,7 @@ The configuration file name for this integration is `python.d/tomcat.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/tomcat/metadata.yaml b/src/collectors/python.d.plugin/tomcat/metadata.yaml
index e68526073..e68526073 100644
--- a/collectors/python.d.plugin/tomcat/metadata.yaml
+++ b/src/collectors/python.d.plugin/tomcat/metadata.yaml
diff --git a/collectors/python.d.plugin/tomcat/tomcat.chart.py b/src/collectors/python.d.plugin/tomcat/tomcat.chart.py
index 90315f8c7..90315f8c7 100644
--- a/collectors/python.d.plugin/tomcat/tomcat.chart.py
+++ b/src/collectors/python.d.plugin/tomcat/tomcat.chart.py
diff --git a/collectors/python.d.plugin/tomcat/tomcat.conf b/src/collectors/python.d.plugin/tomcat/tomcat.conf
index 009591bdf..009591bdf 100644
--- a/collectors/python.d.plugin/tomcat/tomcat.conf
+++ b/src/collectors/python.d.plugin/tomcat/tomcat.conf
diff --git a/collectors/python.d.plugin/tor/README.md b/src/collectors/python.d.plugin/tor/README.md
index 7c20cd40a..7c20cd40a 120000
--- a/collectors/python.d.plugin/tor/README.md
+++ b/src/collectors/python.d.plugin/tor/README.md
diff --git a/collectors/python.d.plugin/tor/integrations/tor.md b/src/collectors/python.d.plugin/tor/integrations/tor.md
index 0e57fa793..cbdf18d16 100644
--- a/collectors/python.d.plugin/tor/integrations/tor.md
+++ b/src/collectors/python.d.plugin/tor/integrations/tor.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/tor/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/tor/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/tor/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/tor/metadata.yaml"
sidebar_label: "Tor"
learn_status: "Published"
-learn_rel_path: "Data Collection/VPNs"
+learn_rel_path: "Collecting Metrics/VPNs"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -99,7 +99,7 @@ The configuration file name for this integration is `python.d/tor.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/tor/metadata.yaml b/src/collectors/python.d.plugin/tor/metadata.yaml
index 8647eca23..8647eca23 100644
--- a/collectors/python.d.plugin/tor/metadata.yaml
+++ b/src/collectors/python.d.plugin/tor/metadata.yaml
diff --git a/collectors/python.d.plugin/tor/tor.chart.py b/src/collectors/python.d.plugin/tor/tor.chart.py
index f7bc2d79b..f7bc2d79b 100644
--- a/collectors/python.d.plugin/tor/tor.chart.py
+++ b/src/collectors/python.d.plugin/tor/tor.chart.py
diff --git a/collectors/python.d.plugin/tor/tor.conf b/src/collectors/python.d.plugin/tor/tor.conf
index c7c98dc0b..c7c98dc0b 100644
--- a/collectors/python.d.plugin/tor/tor.conf
+++ b/src/collectors/python.d.plugin/tor/tor.conf
diff --git a/collectors/python.d.plugin/traefik/traefik.chart.py b/src/collectors/python.d.plugin/traefik/traefik.chart.py
index 5a498467f..5a498467f 100644
--- a/collectors/python.d.plugin/traefik/traefik.chart.py
+++ b/src/collectors/python.d.plugin/traefik/traefik.chart.py
diff --git a/collectors/python.d.plugin/traefik/traefik.conf b/src/collectors/python.d.plugin/traefik/traefik.conf
index e3f182d32..e3f182d32 100644
--- a/collectors/python.d.plugin/traefik/traefik.conf
+++ b/src/collectors/python.d.plugin/traefik/traefik.conf
diff --git a/collectors/python.d.plugin/uwsgi/README.md b/src/collectors/python.d.plugin/uwsgi/README.md
index 44b855949..44b855949 120000
--- a/collectors/python.d.plugin/uwsgi/README.md
+++ b/src/collectors/python.d.plugin/uwsgi/README.md
diff --git a/collectors/python.d.plugin/uwsgi/integrations/uwsgi.md b/src/collectors/python.d.plugin/uwsgi/integrations/uwsgi.md
index af58608bd..e1b27ff1f 100644
--- a/collectors/python.d.plugin/uwsgi/integrations/uwsgi.md
+++ b/src/collectors/python.d.plugin/uwsgi/integrations/uwsgi.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/uwsgi/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/uwsgi/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/uwsgi/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/uwsgi/metadata.yaml"
sidebar_label: "uWSGI"
learn_status: "Published"
-learn_rel_path: "Data Collection/Web Servers and Web Proxies"
+learn_rel_path: "Collecting Metrics/Web Servers and Web Proxies"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -99,7 +99,7 @@ The configuration file name for this integration is `python.d/uwsgi.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/uwsgi/metadata.yaml b/src/collectors/python.d.plugin/uwsgi/metadata.yaml
index cdb090ac1..cdb090ac1 100644
--- a/collectors/python.d.plugin/uwsgi/metadata.yaml
+++ b/src/collectors/python.d.plugin/uwsgi/metadata.yaml
diff --git a/collectors/python.d.plugin/uwsgi/uwsgi.chart.py b/src/collectors/python.d.plugin/uwsgi/uwsgi.chart.py
index e4d900005..e4d900005 100644
--- a/collectors/python.d.plugin/uwsgi/uwsgi.chart.py
+++ b/src/collectors/python.d.plugin/uwsgi/uwsgi.chart.py
diff --git a/collectors/python.d.plugin/uwsgi/uwsgi.conf b/src/collectors/python.d.plugin/uwsgi/uwsgi.conf
index 7d09e7330..7d09e7330 100644
--- a/collectors/python.d.plugin/uwsgi/uwsgi.conf
+++ b/src/collectors/python.d.plugin/uwsgi/uwsgi.conf
diff --git a/collectors/python.d.plugin/varnish/README.md b/src/collectors/python.d.plugin/varnish/README.md
index 194be2335..194be2335 120000
--- a/collectors/python.d.plugin/varnish/README.md
+++ b/src/collectors/python.d.plugin/varnish/README.md
diff --git a/collectors/python.d.plugin/varnish/integrations/varnish.md b/src/collectors/python.d.plugin/varnish/integrations/varnish.md
index da74dcf8f..dd9ce8fe6 100644
--- a/collectors/python.d.plugin/varnish/integrations/varnish.md
+++ b/src/collectors/python.d.plugin/varnish/integrations/varnish.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/varnish/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/varnish/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/varnish/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/varnish/metadata.yaml"
sidebar_label: "Varnish"
learn_status: "Published"
-learn_rel_path: "Data Collection/Web Servers and Web Proxies"
+learn_rel_path: "Collecting Metrics/Web Servers and Web Proxies"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -138,7 +138,7 @@ The configuration file name for this integration is `python.d/varnish.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/varnish/metadata.yaml b/src/collectors/python.d.plugin/varnish/metadata.yaml
index d31c1cf6f..d31c1cf6f 100644
--- a/collectors/python.d.plugin/varnish/metadata.yaml
+++ b/src/collectors/python.d.plugin/varnish/metadata.yaml
diff --git a/collectors/python.d.plugin/varnish/varnish.chart.py b/src/collectors/python.d.plugin/varnish/varnish.chart.py
index 506ad026a..506ad026a 100644
--- a/collectors/python.d.plugin/varnish/varnish.chart.py
+++ b/src/collectors/python.d.plugin/varnish/varnish.chart.py
diff --git a/collectors/python.d.plugin/varnish/varnish.conf b/src/collectors/python.d.plugin/varnish/varnish.conf
index 54bfe4dee..54bfe4dee 100644
--- a/collectors/python.d.plugin/varnish/varnish.conf
+++ b/src/collectors/python.d.plugin/varnish/varnish.conf
diff --git a/collectors/python.d.plugin/w1sensor/README.md b/src/collectors/python.d.plugin/w1sensor/README.md
index c0fa9cd1b..c0fa9cd1b 120000
--- a/collectors/python.d.plugin/w1sensor/README.md
+++ b/src/collectors/python.d.plugin/w1sensor/README.md
diff --git a/collectors/python.d.plugin/w1sensor/integrations/1-wire_sensors.md b/src/collectors/python.d.plugin/w1sensor/integrations/1-wire_sensors.md
index fe3c05ba6..a01dc8195 100644
--- a/collectors/python.d.plugin/w1sensor/integrations/1-wire_sensors.md
+++ b/src/collectors/python.d.plugin/w1sensor/integrations/1-wire_sensors.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/w1sensor/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/w1sensor/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/w1sensor/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/w1sensor/metadata.yaml"
sidebar_label: "1-Wire Sensors"
learn_status: "Published"
-learn_rel_path: "Data Collection/Hardware Devices and Sensors"
+learn_rel_path: "Collecting Metrics/Hardware Devices and Sensors"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -91,7 +91,7 @@ The configuration file name for this integration is `python.d/w1sensor.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/w1sensor/metadata.yaml b/src/collectors/python.d.plugin/w1sensor/metadata.yaml
index 7b0768237..7b0768237 100644
--- a/collectors/python.d.plugin/w1sensor/metadata.yaml
+++ b/src/collectors/python.d.plugin/w1sensor/metadata.yaml
diff --git a/collectors/python.d.plugin/w1sensor/w1sensor.chart.py b/src/collectors/python.d.plugin/w1sensor/w1sensor.chart.py
index 66797ced3..66797ced3 100644
--- a/collectors/python.d.plugin/w1sensor/w1sensor.chart.py
+++ b/src/collectors/python.d.plugin/w1sensor/w1sensor.chart.py
diff --git a/collectors/python.d.plugin/w1sensor/w1sensor.conf b/src/collectors/python.d.plugin/w1sensor/w1sensor.conf
index b60d28650..b60d28650 100644
--- a/collectors/python.d.plugin/w1sensor/w1sensor.conf
+++ b/src/collectors/python.d.plugin/w1sensor/w1sensor.conf
diff --git a/collectors/python.d.plugin/zscores/README.md b/src/collectors/python.d.plugin/zscores/README.md
index 159ce0787..159ce0787 120000
--- a/collectors/python.d.plugin/zscores/README.md
+++ b/src/collectors/python.d.plugin/zscores/README.md
diff --git a/collectors/python.d.plugin/zscores/integrations/python.d_zscores.md b/src/collectors/python.d.plugin/zscores/integrations/python.d_zscores.md
index 9d7d1c3d5..738c359db 100644
--- a/collectors/python.d.plugin/zscores/integrations/python.d_zscores.md
+++ b/src/collectors/python.d.plugin/zscores/integrations/python.d_zscores.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/zscores/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/python.d.plugin/zscores/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/zscores/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/python.d.plugin/zscores/metadata.yaml"
sidebar_label: "python.d zscores"
learn_status: "Published"
-learn_rel_path: "Data Collection/Other"
+learn_rel_path: "Collecting Metrics/Other"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -20,7 +20,7 @@ Module: zscores
By using smoothed, rolling [Z-Scores](https://en.wikipedia.org/wiki/Standard_score) for selected metrics or charts you can narrow down your focus and shorten root cause analysis.
-This collector uses the [Netdata rest api](https://github.com/netdata/netdata/blob/master/web/api/README.md) to get the `mean` and `stddev`
+This collector uses the [Netdata rest api](https://github.com/netdata/netdata/blob/master/src/web/api/README.md) to get the `mean` and `stddev`
for each dimension on specified charts over a time range (defined by `train_secs` and `offset_secs`).
For each dimension it will calculate a Z-Score as `z = (x - mean) / stddev` (clipped at `z_clip`). Scores are then smoothed over
@@ -100,7 +100,7 @@ The configuration file name for this integration is `python.d/zscores.conf`.
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/python.d.plugin/zscores/zscores.chart.py b/src/collectors/python.d.plugin/zscores/zscores.chart.py
index 1099b9376..1099b9376 100644
--- a/collectors/python.d.plugin/zscores/zscores.chart.py
+++ b/src/collectors/python.d.plugin/zscores/zscores.chart.py
diff --git a/collectors/python.d.plugin/zscores/zscores.conf b/src/collectors/python.d.plugin/zscores/zscores.conf
index 07d62ebe6..07d62ebe6 100644
--- a/collectors/python.d.plugin/zscores/zscores.conf
+++ b/src/collectors/python.d.plugin/zscores/zscores.conf
diff --git a/collectors/slabinfo.plugin/README.md b/src/collectors/slabinfo.plugin/README.md
index 4d4629a77..4d4629a77 120000
--- a/collectors/slabinfo.plugin/README.md
+++ b/src/collectors/slabinfo.plugin/README.md
diff --git a/collectors/slabinfo.plugin/integrations/linux_kernel_slab_allocator_statistics.md b/src/collectors/slabinfo.plugin/integrations/linux_kernel_slab_allocator_statistics.md
index ce8115270..858fcba46 100644
--- a/collectors/slabinfo.plugin/integrations/linux_kernel_slab_allocator_statistics.md
+++ b/src/collectors/slabinfo.plugin/integrations/linux_kernel_slab_allocator_statistics.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/slabinfo.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/slabinfo.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/slabinfo.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/slabinfo.plugin/metadata.yaml"
sidebar_label: "Linux kernel SLAB allocator statistics"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Kernel"
+learn_rel_path: "Collecting Metrics/Linux Systems/Kernel"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -107,7 +107,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/slabinfo.plugin/metadata.yaml b/src/collectors/slabinfo.plugin/metadata.yaml
index f19778297..f19778297 100644
--- a/collectors/slabinfo.plugin/metadata.yaml
+++ b/src/collectors/slabinfo.plugin/metadata.yaml
diff --git a/collectors/slabinfo.plugin/slabinfo.c b/src/collectors/slabinfo.plugin/slabinfo.c
index 9b9119a6e..58a34ad93 100644
--- a/collectors/slabinfo.plugin/slabinfo.c
+++ b/src/collectors/slabinfo.plugin/slabinfo.c
@@ -11,10 +11,13 @@
#define CHART_PRIO 3000
// #define slabdebug(...) if (debug) { fprintf(stderr, __VA_ARGS__); }
-#define slabdebug(args...) if (debug) { \
- fprintf(stderr, "slabinfo.plugin DEBUG (%04d@%-10.10s:%-15.15s)::", __LINE__, __FILE__, __FUNCTION__); \
- fprintf(stderr, ##args); \
- fprintf(stderr, "\n"); }
+#define slabdebug(args...) do { \
+ if (debug) { \
+ fprintf(stderr, "slabinfo.plugin DEBUG (%04d@%-10.10s:%-15.15s)::", __LINE__, __FILE__, __FUNCTION__); \
+ fprintf(stderr, ##args); \
+ fprintf(stderr, "\n"); \
+ } \
+} while(0)
int running = 1;
int debug = 0;
diff --git a/collectors/statsd.plugin/asterisk.conf b/src/collectors/statsd.plugin/asterisk.conf
index 160b80f93..160b80f93 100644
--- a/collectors/statsd.plugin/asterisk.conf
+++ b/src/collectors/statsd.plugin/asterisk.conf
diff --git a/collectors/statsd.plugin/asterisk.md b/src/collectors/statsd.plugin/asterisk.md
index e7a7b63ce..302fb932f 100644
--- a/collectors/statsd.plugin/asterisk.md
+++ b/src/collectors/statsd.plugin/asterisk.md
@@ -1,6 +1,6 @@
<!--
title: "Asterisk monitoring with Netdata"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/statsd.plugin/asterisk.md"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/statsd.plugin/asterisk.md"
sidebar_label: "Asterisk"
learn_status: "Published"
learn_rel_path: "Integrations/Monitor/VoIP"
@@ -17,7 +17,7 @@ Monitors [Asterisk](https://www.asterisk.org/) dialplan application's statistics
## Configuration
Netdata ships
-with [asterisk.conf](https://github.com/netdata/netdata/blob/master/collectors/statsd.plugin/asterisk.conf) with
+with [asterisk.conf](https://github.com/netdata/netdata/blob/master/src/collectors/statsd.plugin/asterisk.conf) with
preconfigured charts.
To receive Asterisk metrics in Netdata, uncomment the following lines in the `/etc/asterisk/statsd.conf` file:
diff --git a/collectors/statsd.plugin/example.conf b/src/collectors/statsd.plugin/example.conf
index 2c7de6c7b..2c7de6c7b 100644
--- a/collectors/statsd.plugin/example.conf
+++ b/src/collectors/statsd.plugin/example.conf
diff --git a/collectors/statsd.plugin/k6.conf b/src/collectors/statsd.plugin/k6.conf
index 3bef00ca1..3bef00ca1 100644
--- a/collectors/statsd.plugin/k6.conf
+++ b/src/collectors/statsd.plugin/k6.conf
diff --git a/collectors/statsd.plugin/k6.md b/src/collectors/statsd.plugin/k6.md
index 13608a8a8..b657ff1a9 100644
--- a/collectors/statsd.plugin/k6.md
+++ b/src/collectors/statsd.plugin/k6.md
@@ -1,6 +1,6 @@
<!--
title: "K6 load test monitoring with Netdata"
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/statsd.plugin/k6.md"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/statsd.plugin/k6.md"
sidebar_label: "K6 Load Testing"
learn_status: "Published"
learn_rel_path: "Integrations/Monitor/apps"
diff --git a/collectors/statsd.plugin/statsd.c b/src/collectors/statsd.plugin/statsd.c
index 9cc3a9d97..b892133d2 100644
--- a/collectors/statsd.plugin/statsd.c
+++ b/src/collectors/statsd.plugin/statsd.c
@@ -1531,8 +1531,7 @@ static int statsd_readfile(const char *filename, STATSD_APP *app, STATSD_APP_CHA
return 0;
}
-static int statsd_file_callback(const char *filename, void *data) {
- (void)data;
+static int statsd_file_callback(const char *filename, void *data __maybe_unused, bool stock_config __maybe_unused) {
return statsd_readfile(filename, NULL, NULL, NULL);
}
diff --git a/collectors/systemd-journal.plugin/forward_secure_sealing.md b/src/collectors/systemd-journal.plugin/forward_secure_sealing.md
index b41570d68..b41570d68 100644
--- a/collectors/systemd-journal.plugin/forward_secure_sealing.md
+++ b/src/collectors/systemd-journal.plugin/forward_secure_sealing.md
diff --git a/collectors/systemd-journal.plugin/passive_journal_centralization_guide_self_signed_certs.md b/src/collectors/systemd-journal.plugin/passive_journal_centralization_guide_self_signed_certs.md
index 722d1ceae..f8b9a62f0 100644
--- a/collectors/systemd-journal.plugin/passive_journal_centralization_guide_self_signed_certs.md
+++ b/src/collectors/systemd-journal.plugin/passive_journal_centralization_guide_self_signed_certs.md
@@ -20,7 +20,7 @@ This means, that if both certificates are issued by the same certificate authori
## Self-signed certificates
-To simplify the process of creating and managing self-signed certificates, we have created [this bash script](https://github.com/netdata/netdata/blob/master/collectors/systemd-journal.plugin/systemd-journal-self-signed-certs.sh).
+To simplify the process of creating and managing self-signed certificates, we have created [this bash script](https://github.com/netdata/netdata/blob/master/src/collectors/systemd-journal.plugin/systemd-journal-self-signed-certs.sh).
This helps to also automate the distribution of the certificates to your servers (it generates a new bash script for each of your servers, which includes everything required, including the certificates).
@@ -34,7 +34,7 @@ On the server that will issue the certificates (usually the centralizaton server
sudo apt-get install systemd-journal-remote openssl
# download the script and make it executable
-curl >systemd-journal-self-signed-certs.sh "https://raw.githubusercontent.com/netdata/netdata/master/collectors/systemd-journal.plugin/systemd-journal-self-signed-certs.sh"
+curl >systemd-journal-self-signed-certs.sh "https://raw.githubusercontent.com/netdata/netdata/master/src/collectors/systemd-journal.plugin/systemd-journal-self-signed-certs.sh"
chmod 750 systemd-journal-self-signed-certs.sh
```
diff --git a/collectors/systemd-journal.plugin/systemd-internals.h b/src/collectors/systemd-journal.plugin/systemd-internals.h
index e1ae44d4f..31acb2f20 100644
--- a/collectors/systemd-journal.plugin/systemd-internals.h
+++ b/src/collectors/systemd-journal.plugin/systemd-internals.h
@@ -83,12 +83,11 @@ struct journal_file {
#define ND_SD_JOURNAL_OPEN_FLAGS (0)
-#define JOURNAL_VS_REALTIME_DELTA_DEFAULT_UT (5 * USEC_PER_SEC) // assume always 5 seconds latency
-#define JOURNAL_VS_REALTIME_DELTA_MAX_UT (2 * 60 * USEC_PER_SEC) // up to 2 minutes latency
+#define JOURNAL_VS_REALTIME_DELTA_DEFAULT_UT (5 * USEC_PER_SEC) // assume a 5-seconds latency
+#define JOURNAL_VS_REALTIME_DELTA_MAX_UT (2 * 60 * USEC_PER_SEC) // up to 2-minutes latency
extern DICTIONARY *journal_files_registry;
extern DICTIONARY *used_hashes_registry;
-extern DICTIONARY *function_query_status_dict;
extern DICTIONARY *boot_ids_to_first_ut;
int journal_file_dict_items_backward_compar(const void *a, const void *b);
@@ -115,31 +114,24 @@ usec_t journal_file_update_annotation_boot_id(sd_journal *j, struct journal_file
#define MAX_JOURNAL_DIRECTORIES 100
struct journal_directory {
- char *path;
+ STRING *path;
};
extern struct journal_directory journal_directories[MAX_JOURNAL_DIRECTORIES];
void journal_init_files_and_directories(void);
-void journal_init_query_status(void);
-void function_systemd_journal(const char *transaction, char *function, int timeout, bool *cancelled);
+void function_systemd_journal(const char *transaction, char *function, usec_t *stop_monotonic_ut, bool *cancelled, BUFFER *payload, HTTP_ACCESS access __maybe_unused, const char *source, void *data);
void journal_file_update_header(const char *filename, struct journal_file *jf);
-void netdata_systemd_journal_message_ids_init(void);
-void netdata_systemd_journal_transform_message_id(FACETS *facets __maybe_unused, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope __maybe_unused, void *data __maybe_unused);
+void netdata_systemd_journal_annotations_init(void);
+void netdata_systemd_journal_transform_message_id(FACETS *facets, BUFFER *wb, FACETS_TRANSFORMATION_SCOPE scope, void *data);
void *journal_watcher_main(void *arg);
+void journal_watcher_restart(void);
#ifdef ENABLE_SYSTEMD_DBUS
-void function_systemd_units(const char *transaction, char *function, int timeout, bool *cancelled);
+void function_systemd_units(const char *transaction, char *function, usec_t *stop_monotonic_ut, bool *cancelled, BUFFER *payload, HTTP_ACCESS access __maybe_unused, const char *source, void *data);
#endif
-static inline void send_newline_and_flush(void) {
- netdata_mutex_lock(&stdout_mutex);
- fprintf(stdout, "\n");
- fflush(stdout);
- netdata_mutex_unlock(&stdout_mutex);
-}
-
static inline bool parse_journal_field(const char *data, size_t data_length, const char **key, size_t *key_length, const char **value, size_t *value_length) {
const char *k = data;
const char *equal = strchr(k, '=');
@@ -159,4 +151,6 @@ static inline bool parse_journal_field(const char *data, size_t data_length, con
return true;
}
+void systemd_journal_dyncfg_init(struct functions_evloop_globals *wg);
+
#endif //NETDATA_COLLECTORS_SYSTEMD_INTERNALS_H
diff --git a/collectors/systemd-journal.plugin/systemd-journal-files.c b/src/collectors/systemd-journal.plugin/systemd-journal-files.c
index 56496df22..a05cd1c5c 100644
--- a/collectors/systemd-journal.plugin/systemd-journal-files.c
+++ b/src/collectors/systemd-journal.plugin/systemd-journal-files.c
@@ -39,74 +39,74 @@ static bool journal_sd_id128_parse(const char *in, sd_id128_t *ret) {
return false;
}
-static void journal_file_get_header_from_journalctl(const char *filename, struct journal_file *jf) {
- // unfortunately, our capabilities are not inheritted by journalctl
- // so, it fails to give us the information we need.
-
- bool read_writer = false, read_head = false, read_tail = false;
-
- char cmd[FILENAME_MAX * 2];
- snprintfz(cmd, sizeof(cmd), "journalctl --header --file '%s'", filename);
- CLEAN_BUFFER *wb = run_command_and_get_output_to_buffer(cmd, 1024);
- if(wb) {
- const char *s = buffer_tostring(wb);
-
- const char *sequential_id_header = "Sequential Number ID:";
- const char *sequential_id_data = strcasestr(s, sequential_id_header);
- if(sequential_id_data) {
- sequential_id_data += strlen(sequential_id_header);
- if(journal_sd_id128_parse(sequential_id_data, &jf->first_writer_id))
- read_writer = true;
- }
-
- const char *head_sequential_number_header = "Head sequential number:";
- const char *head_sequential_number_data = strcasestr(s, head_sequential_number_header);
- if(head_sequential_number_data) {
- head_sequential_number_data += strlen(head_sequential_number_header);
-
- while(isspace(*head_sequential_number_data))
- head_sequential_number_data++;
-
- if(isdigit(*head_sequential_number_data)) {
- jf->first_seqnum = strtoul(head_sequential_number_data, NULL, 10);
- if(jf->first_seqnum)
- read_head = true;
- }
- }
-
- const char *tail_sequential_number_header = "Tail sequential number:";
- const char *tail_sequential_number_data = strcasestr(s, tail_sequential_number_header);
- if(tail_sequential_number_data) {
- tail_sequential_number_data += strlen(tail_sequential_number_header);
-
- while(isspace(*tail_sequential_number_data))
- tail_sequential_number_data++;
-
- if(isdigit(*tail_sequential_number_data)) {
- jf->last_seqnum = strtoul(tail_sequential_number_data, NULL, 10);
- if(jf->last_seqnum)
- read_tail = true;
- }
- }
-
- if(read_head && read_tail && jf->last_seqnum > jf->first_seqnum)
- jf->messages_in_file = jf->last_seqnum - jf->first_seqnum;
- }
-
- if(!jf->logged_journalctl_failure && (!read_head || !read_head || !read_tail)) {
-
- nd_log(NDLS_COLLECTORS, NDLP_NOTICE,
- "Failed to read %s%s%s from journalctl's output on filename '%s', using the command: %s",
- read_writer?"":"writer id,",
- read_head?"":"head id,",
- read_tail?"":"tail id,",
- filename, cmd);
-
- jf->logged_journalctl_failure = true;
- }
-}
-
-usec_t journal_file_update_annotation_boot_id(sd_journal *j, struct journal_file *jf, const char *boot_id) {
+//static void journal_file_get_header_from_journalctl(const char *filename, struct journal_file *jf) {
+// // unfortunately, our capabilities are not inheritted by journalctl
+// // so, it fails to give us the information we need.
+//
+// bool read_writer = false, read_head = false, read_tail = false;
+//
+// char cmd[FILENAME_MAX * 2];
+// snprintfz(cmd, sizeof(cmd), "journalctl --header --file '%s'", filename);
+// CLEAN_BUFFER *wb = run_command_and_get_output_to_buffer(cmd, 1024);
+// if(wb) {
+// const char *s = buffer_tostring(wb);
+//
+// const char *sequential_id_header = "Sequential Number ID:";
+// const char *sequential_id_data = strcasestr(s, sequential_id_header);
+// if(sequential_id_data) {
+// sequential_id_data += strlen(sequential_id_header);
+// if(journal_sd_id128_parse(sequential_id_data, &jf->first_writer_id))
+// read_writer = true;
+// }
+//
+// const char *head_sequential_number_header = "Head sequential number:";
+// const char *head_sequential_number_data = strcasestr(s, head_sequential_number_header);
+// if(head_sequential_number_data) {
+// head_sequential_number_data += strlen(head_sequential_number_header);
+//
+// while(isspace(*head_sequential_number_data))
+// head_sequential_number_data++;
+//
+// if(isdigit(*head_sequential_number_data)) {
+// jf->first_seqnum = strtoul(head_sequential_number_data, NULL, 10);
+// if(jf->first_seqnum)
+// read_head = true;
+// }
+// }
+//
+// const char *tail_sequential_number_header = "Tail sequential number:";
+// const char *tail_sequential_number_data = strcasestr(s, tail_sequential_number_header);
+// if(tail_sequential_number_data) {
+// tail_sequential_number_data += strlen(tail_sequential_number_header);
+//
+// while(isspace(*tail_sequential_number_data))
+// tail_sequential_number_data++;
+//
+// if(isdigit(*tail_sequential_number_data)) {
+// jf->last_seqnum = strtoul(tail_sequential_number_data, NULL, 10);
+// if(jf->last_seqnum)
+// read_tail = true;
+// }
+// }
+//
+// if(read_head && read_tail && jf->last_seqnum > jf->first_seqnum)
+// jf->messages_in_file = jf->last_seqnum - jf->first_seqnum;
+// }
+//
+// if(!jf->logged_journalctl_failure && (!read_head || !read_tail)) {
+//
+// nd_log(NDLS_COLLECTORS, NDLP_NOTICE,
+// "Failed to read %s%s%s from journalctl's output on filename '%s', using the command: %s",
+// read_writer?"":"writer id,",
+// read_head?"":"head id,",
+// read_tail?"":"tail id,",
+// filename, cmd);
+//
+// jf->logged_journalctl_failure = true;
+// }
+//}
+
+usec_t journal_file_update_annotation_boot_id(sd_journal *j, struct journal_file *jf __maybe_unused, const char *boot_id) {
usec_t ut = UINT64_MAX;
int r;
@@ -447,7 +447,7 @@ static void files_registry_insert_cb(const DICTIONARY_ITEM *item, void *value, v
jf->filename);
}
-static bool files_registry_conflict_cb(const DICTIONARY_ITEM *item, void *old_value, void *new_value, void *data __maybe_unused) {
+static bool files_registry_conflict_cb(const DICTIONARY_ITEM *item __maybe_unused, void *old_value, void *new_value, void *data __maybe_unused) {
struct journal_file *jf = old_value;
struct journal_file *njf = new_value;
@@ -639,7 +639,7 @@ void journal_directory_scan_recursively(DICTIONARY *files, DICTIONARY *dirs, con
if(files)
dictionary_set(files, full_path, NULL, 0);
- send_newline_and_flush();
+ send_newline_and_flush(&stdout_mutex);
}
else if (entry->d_type == DT_LNK) {
struct stat info;
@@ -657,7 +657,7 @@ void journal_directory_scan_recursively(DICTIONARY *files, DICTIONARY *dirs, con
if(files)
dictionary_set(files, full_path, NULL, 0);
- send_newline_and_flush();
+ send_newline_and_flush(&stdout_mutex);
}
}
}
@@ -715,7 +715,7 @@ void journal_files_registry_update(void) {
for(unsigned i = 0; i < MAX_JOURNAL_DIRECTORIES; i++) {
if(!journal_directories[i].path) break;
- journal_directory_scan_recursively(files, dirs, journal_directories[i].path, 0);
+ journal_directory_scan_recursively(files, dirs, string2str(journal_directories[i].path), 0);
}
const char **array = mallocz(sizeof(const char *) * dictionary_entries(files));
@@ -801,7 +801,7 @@ int journal_file_dict_items_forward_compar(const void *a, const void *b) {
return -journal_file_dict_items_backward_compar(a, b);
}
-static bool boot_id_conflict_cb(const DICTIONARY_ITEM *item, void *old_value, void *new_value, void *data __maybe_unused) {
+static bool boot_id_conflict_cb(const DICTIONARY_ITEM *item __maybe_unused, void *old_value, void *new_value, void *data __maybe_unused) {
usec_t *old_usec = old_value;
usec_t *new_usec = new_value;
@@ -819,15 +819,15 @@ void journal_init_files_and_directories(void) {
// ------------------------------------------------------------------------
// setup the journal directories
- journal_directories[d++].path = strdupz("/run/log/journal");
- journal_directories[d++].path = strdupz("/var/log/journal");
+ journal_directories[d++].path = string_strdupz("/run/log/journal");
+ journal_directories[d++].path = string_strdupz("/var/log/journal");
if(*netdata_configured_host_prefix) {
char path[PATH_MAX];
snprintfz(path, sizeof(path), "%s/var/log/journal", netdata_configured_host_prefix);
- journal_directories[d++].path = strdupz(path);
+ journal_directories[d++].path = string_strdupz(path);
snprintfz(path, sizeof(path), "%s/run/log/journal", netdata_configured_host_prefix);
- journal_directories[d++].path = strdupz(path);
+ journal_directories[d++].path = string_strdupz(path);
}
// terminate the list
diff --git a/collectors/systemd-journal.plugin/systemd-journal-fstat.c b/src/collectors/systemd-journal.plugin/systemd-journal-fstat.c
index 45ea78174..45ea78174 100644
--- a/collectors/systemd-journal.plugin/systemd-journal-fstat.c
+++ b/src/collectors/systemd-journal.plugin/systemd-journal-fstat.c
diff --git a/collectors/systemd-journal.plugin/systemd-journal-self-signed-certs.sh b/src/collectors/systemd-journal.plugin/systemd-journal-self-signed-certs.sh
index ada735f1f..ada735f1f 100755
--- a/collectors/systemd-journal.plugin/systemd-journal-self-signed-certs.sh
+++ b/src/collectors/systemd-journal.plugin/systemd-journal-self-signed-certs.sh
diff --git a/collectors/systemd-journal.plugin/systemd-journal-watcher.c b/src/collectors/systemd-journal.plugin/systemd-journal-watcher.c
index ed41f6247..6f12f154e 100644
--- a/collectors/systemd-journal.plugin/systemd-journal-watcher.c
+++ b/src/collectors/systemd-journal.plugin/systemd-journal-watcher.c
@@ -292,8 +292,16 @@ static void process_pending(Watcher *watcher) {
dictionary_garbage_collect(watcher->pending);
}
+size_t journal_watcher_wanted_session_id = 0;
+
+void journal_watcher_restart(void) {
+ __atomic_add_fetch(&journal_watcher_wanted_session_id, 1, __ATOMIC_RELAXED);
+}
+
void *journal_watcher_main(void *arg __maybe_unused) {
while(1) {
+ size_t journal_watcher_session_id = __atomic_load_n(&journal_watcher_wanted_session_id, __ATOMIC_RELAXED);
+
Watcher watcher = {
.watchList = mallocz(INITIAL_WATCHES * sizeof(WatchEntry)),
.freeList = NULL,
@@ -312,12 +320,12 @@ void *journal_watcher_main(void *arg __maybe_unused) {
for (unsigned i = 0; i < MAX_JOURNAL_DIRECTORIES; i++) {
if (!journal_directories[i].path) break;
- watch_directory_and_subdirectories(&watcher, inotifyFd, journal_directories[i].path);
+ watch_directory_and_subdirectories(&watcher, inotifyFd, string2str(journal_directories[i].path));
}
usec_t last_headers_update_ut = now_monotonic_usec();
struct buffered_reader reader;
- while (1) {
+ while (journal_watcher_session_id == __atomic_load_n(&journal_watcher_wanted_session_id, __ATOMIC_RELAXED)) {
buffered_reader_ret_t rc = buffered_reader_read_timeout(
&reader, inotifyFd, SYSTEMD_JOURNAL_EXECUTE_WATCHER_PENDING_EVERY_MS, false);
@@ -372,7 +380,7 @@ void *journal_watcher_main(void *arg __maybe_unused) {
// this will scan the directories and cleanup the registry
journal_files_registry_update();
- sleep_usec(5 * USEC_PER_SEC);
+ sleep_usec(2 * USEC_PER_SEC);
}
return NULL;
diff --git a/collectors/systemd-journal.plugin/systemd-journal.c b/src/collectors/systemd-journal.plugin/systemd-journal.c
index f812b2161..84f644575 100644
--- a/collectors/systemd-journal.plugin/systemd-journal.c
+++ b/src/collectors/systemd-journal.plugin/systemd-journal.c
@@ -26,6 +26,8 @@
#define SYSTEMD_JOURNAL_SAMPLING_SLOTS 1000
#define SYSTEMD_JOURNAL_SAMPLING_RECALIBRATE 10000
+#define SYSTEMD_JOURNAL_PROGRESS_EVERY_UT (250 * USEC_PER_MS)
+
#define JOURNAL_PARAMETER_HELP "help"
#define JOURNAL_PARAMETER_AFTER "after"
#define JOURNAL_PARAMETER_BEFORE "before"
@@ -39,8 +41,6 @@
#define JOURNAL_PARAMETER_DATA_ONLY "data_only"
#define JOURNAL_PARAMETER_SOURCE "source"
#define JOURNAL_PARAMETER_INFO "info"
-#define JOURNAL_PARAMETER_ID "id"
-#define JOURNAL_PARAMETER_PROGRESS "progress"
#define JOURNAL_PARAMETER_SLICE "slice"
#define JOURNAL_PARAMETER_DELTA "delta"
#define JOURNAL_PARAMETER_TAIL "tail"
@@ -183,11 +183,11 @@
typedef struct function_query_status {
bool *cancelled; // a pointer to the cancelling boolean
- usec_t stop_monotonic_ut;
-
- usec_t started_monotonic_ut;
+ usec_t *stop_monotonic_ut;
// request
+ const char *transaction;
+
SD_JOURNAL_FILE_SOURCE_TYPE source_type;
SIMPLE_PATTERN *sources;
usec_t after_ut;
@@ -518,12 +518,12 @@ static size_t sampling_running_file_query_estimate_remaining_lines_by_time(FUNCT
return remaining_logs_by_time;
}
-static size_t sampling_running_file_query_estimate_remaining_lines(sd_journal *j, FUNCTION_QUERY_STATUS *fqs, struct journal_file *jf, FACETS_ANCHOR_DIRECTION direction, usec_t msg_ut) {
- size_t expected_matching_logs_by_seqnum = 0;
- double proportion_by_seqnum = 0.0;
+static size_t sampling_running_file_query_estimate_remaining_lines(sd_journal *j __maybe_unused, FUNCTION_QUERY_STATUS *fqs, struct journal_file *jf, FACETS_ANCHOR_DIRECTION direction, usec_t msg_ut) {
size_t remaining_logs_by_seqnum = 0;
#ifdef HAVE_SD_JOURNAL_GET_SEQNUM
+ size_t expected_matching_logs_by_seqnum = 0;
+ double proportion_by_seqnum = 0.0;
uint64_t current_msg_seqnum;
sd_id128_t current_msg_writer;
if(!fqs->query_file.first_msg_seqnum || sd_journal_get_seqnum(j, &current_msg_seqnum, &current_msg_writer) < 0) {
@@ -807,7 +807,7 @@ ND_SD_JOURNAL_STATUS netdata_systemd_journal_query_backward(
FUNCTION_PROGRESS_UPDATE_BYTES(fqs->bytes_read, bytes - last_bytes);
last_bytes = bytes;
- status = check_stop(fqs->cancelled, &fqs->stop_monotonic_ut);
+ status = check_stop(fqs->cancelled, fqs->stop_monotonic_ut);
}
}
else if(sample == SAMPLING_SKIP_FIELDS)
@@ -914,7 +914,7 @@ ND_SD_JOURNAL_STATUS netdata_systemd_journal_query_forward(
FUNCTION_PROGRESS_UPDATE_BYTES(fqs->bytes_read, bytes - last_bytes);
last_bytes = bytes;
- status = check_stop(fqs->cancelled, &fqs->stop_monotonic_ut);
+ status = check_stop(fqs->cancelled, fqs->stop_monotonic_ut);
}
}
else if(sample == SAMPLING_SKIP_FIELDS)
@@ -1085,7 +1085,7 @@ static bool jf_is_mine(struct journal_file *jf, FUNCTION_QUERY_STATUS *fqs) {
if((fqs->source_type == SDJF_NONE && !fqs->sources) || (jf->source_type & fqs->source_type) ||
(fqs->sources && simple_pattern_matches(fqs->sources, string2str(jf->source)))) {
- if(!jf->msg_last_ut || !jf->msg_last_ut)
+ if(!jf->msg_last_ut)
// the file is not scanned yet, or the timestamps have not been updated,
// so we don't know if it can contribute or not - let's add it.
return true;
@@ -1150,6 +1150,7 @@ static int netdata_systemd_journal_query(BUFFER *wb, FACETS *facets, FUNCTION_QU
usec_t started_ut = query_started_ut;
usec_t ended_ut = started_ut;
usec_t duration_ut = 0, max_duration_ut = 0;
+ usec_t progress_duration_ut = 0;
sampling_query_init(fqs, facets);
@@ -1164,9 +1165,7 @@ static int netdata_systemd_journal_query(BUFFER *wb, FACETS *facets, FUNCTION_QU
started_ut = ended_ut;
// do not even try to do the query if we expect it to pass the timeout
- if(ended_ut > (query_started_ut + (fqs->stop_monotonic_ut - query_started_ut) * 3 / 4) &&
- ended_ut + max_duration_ut * 2 >= fqs->stop_monotonic_ut) {
-
+ if(ended_ut + max_duration_ut * 3 >= *fqs->stop_monotonic_ut) {
partial = true;
status = ND_SD_JOURNAL_TIMED_OUT;
break;
@@ -1211,6 +1210,14 @@ static int netdata_systemd_journal_query(BUFFER *wb, FACETS *facets, FUNCTION_QU
if(duration_ut > max_duration_ut)
max_duration_ut = duration_ut;
+ progress_duration_ut += duration_ut;
+ if(progress_duration_ut >= SYSTEMD_JOURNAL_PROGRESS_EVERY_UT) {
+ progress_duration_ut = 0;
+ netdata_mutex_lock(&stdout_mutex);
+ pluginsd_function_progress_to_stdout(fqs->transaction, f + 1, files_used);
+ netdata_mutex_unlock(&stdout_mutex);
+ }
+
buffer_json_add_array_item_object(wb); // journal file
{
// information about the file
@@ -1422,15 +1429,6 @@ static void netdata_systemd_journal_function_help(const char *transaction) {
" all the available systemd journal sources.\n"
" When `"JOURNAL_PARAMETER_INFO"` is requested, all other parameters are ignored.\n"
"\n"
- " "JOURNAL_PARAMETER_ID":STRING\n"
- " Caller supplied unique ID of the request.\n"
- " This can be used later to request a progress report of the query.\n"
- " Optional, but if omitted no `"JOURNAL_PARAMETER_PROGRESS"` can be requested.\n"
- "\n"
- " "JOURNAL_PARAMETER_PROGRESS"\n"
- " Request a progress report (the `id` of a running query is required).\n"
- " When `"JOURNAL_PARAMETER_PROGRESS"` is requested, only parameter `"JOURNAL_PARAMETER_ID"` is used.\n"
- "\n"
" "JOURNAL_PARAMETER_DATA_ONLY":true or "JOURNAL_PARAMETER_DATA_ONLY":false\n"
" Quickly respond with data requested, without generating a\n"
" `histogram`, `facets` counters and `items`.\n"
@@ -1522,67 +1520,9 @@ static void netdata_systemd_journal_function_help(const char *transaction) {
buffer_free(wb);
}
-DICTIONARY *function_query_status_dict = NULL;
-
-static void function_systemd_journal_progress(BUFFER *wb, const char *transaction, const char *progress_id) {
- if(!progress_id || !(*progress_id)) {
- netdata_mutex_lock(&stdout_mutex);
- pluginsd_function_json_error_to_stdout(transaction, HTTP_RESP_BAD_REQUEST, "missing progress id");
- netdata_mutex_unlock(&stdout_mutex);
- return;
- }
-
- const DICTIONARY_ITEM *item = dictionary_get_and_acquire_item(function_query_status_dict, progress_id);
-
- if(!item) {
- netdata_mutex_lock(&stdout_mutex);
- pluginsd_function_json_error_to_stdout(transaction, HTTP_RESP_NOT_FOUND, "progress id is not found here");
- netdata_mutex_unlock(&stdout_mutex);
- return;
- }
-
- FUNCTION_QUERY_STATUS *fqs = dictionary_acquired_item_value(item);
-
- usec_t now_monotonic_ut = now_monotonic_usec();
- if(now_monotonic_ut + 10 * USEC_PER_SEC > fqs->stop_monotonic_ut)
- fqs->stop_monotonic_ut = now_monotonic_ut + 10 * USEC_PER_SEC;
-
- usec_t duration_ut = now_monotonic_ut - fqs->started_monotonic_ut;
-
- size_t files_matched = fqs->files_matched;
- size_t file_working = fqs->file_working;
- if(file_working > files_matched)
- files_matched = file_working;
-
- size_t rows_read = __atomic_load_n(&fqs->rows_read, __ATOMIC_RELAXED);
- size_t bytes_read = __atomic_load_n(&fqs->bytes_read, __ATOMIC_RELAXED);
-
- buffer_flush(wb);
- buffer_json_initialize(wb, "\"", "\"", 0, true, BUFFER_JSON_OPTIONS_MINIFY);
- buffer_json_member_add_uint64(wb, "status", HTTP_RESP_OK);
- buffer_json_member_add_string(wb, "type", "table");
- buffer_json_member_add_uint64(wb, "running_duration_usec", duration_ut);
- buffer_json_member_add_double(wb, "progress", (double)file_working * 100.0 / (double)files_matched);
- char msg[1024 + 1];
- snprintfz(msg, sizeof(msg) - 1,
- "Read %zu rows (%0.0f rows/s), "
- "data %0.1f MB (%0.1f MB/s), "
- "file %zu of %zu",
- rows_read, (double)rows_read / (double)duration_ut * (double)USEC_PER_SEC,
- (double)bytes_read / 1024.0 / 1024.0, ((double)bytes_read / (double)duration_ut * (double)USEC_PER_SEC) / 1024.0 / 1024.0,
- file_working, files_matched
- );
- buffer_json_member_add_string(wb, "message", msg);
- buffer_json_finalize(wb);
-
- netdata_mutex_lock(&stdout_mutex);
- pluginsd_function_result_to_stdout(transaction, HTTP_RESP_OK, "application/json", now_realtime_sec() + 1, wb);
- netdata_mutex_unlock(&stdout_mutex);
-
- dictionary_acquired_item_release(function_query_status_dict, item);
-}
-
-void function_systemd_journal(const char *transaction, char *function, int timeout, bool *cancelled) {
+void function_systemd_journal(const char *transaction, char *function, usec_t *stop_monotonic_ut, bool *cancelled,
+ BUFFER *payload __maybe_unused, HTTP_ACCESS access __maybe_unused,
+ const char *source __maybe_unused, void *data __maybe_unused) {
fstat_thread_calls = 0;
fstat_thread_cached_responses = 0;
@@ -1590,14 +1530,11 @@ void function_systemd_journal(const char *transaction, char *function, int timeo
buffer_flush(wb);
buffer_json_initialize(wb, "\"", "\"", 0, true, BUFFER_JSON_OPTIONS_MINIFY);
- usec_t now_monotonic_ut = now_monotonic_usec();
FUNCTION_QUERY_STATUS tmp_fqs = {
.cancelled = cancelled,
- .started_monotonic_ut = now_monotonic_ut,
- .stop_monotonic_ut = now_monotonic_ut + (timeout * USEC_PER_SEC),
+ .stop_monotonic_ut = stop_monotonic_ut,
};
FUNCTION_QUERY_STATUS *fqs = NULL;
- const DICTIONARY_ITEM *fqs_item = NULL;
FACETS *facets = facets_create(50, FACETS_OPTION_ALL_KEYS_FTS,
SYSTEMD_ALWAYS_VISIBLE_KEYS,
@@ -1616,8 +1553,6 @@ void function_systemd_journal(const char *transaction, char *function, int timeo
facets_accepted_param(facets, JOURNAL_PARAMETER_HISTOGRAM);
facets_accepted_param(facets, JOURNAL_PARAMETER_IF_MODIFIED_SINCE);
facets_accepted_param(facets, JOURNAL_PARAMETER_DATA_ONLY);
- facets_accepted_param(facets, JOURNAL_PARAMETER_ID);
- facets_accepted_param(facets, JOURNAL_PARAMETER_PROGRESS);
facets_accepted_param(facets, JOURNAL_PARAMETER_DELTA);
facets_accepted_param(facets, JOURNAL_PARAMETER_TAIL);
facets_accepted_param(facets, JOURNAL_PARAMETER_SAMPLING);
@@ -1724,7 +1659,7 @@ void function_systemd_journal(const char *transaction, char *function, int timeo
// ------------------------------------------------------------------------
// parse the parameters
- bool info = false, data_only = false, progress = false, slice = JOURNAL_DEFAULT_SLICE_MODE, delta = false, tail = false;
+ bool info = false, data_only = false, slice = JOURNAL_DEFAULT_SLICE_MODE, delta = false, tail = false;
time_t after_s = 0, before_s = 0;
usec_t anchor = 0;
usec_t if_modified_since = 0;
@@ -1733,7 +1668,6 @@ void function_systemd_journal(const char *transaction, char *function, int timeo
const char *query = NULL;
const char *chart = NULL;
SIMPLE_PATTERN *sources = NULL;
- const char *progress_id = NULL;
SD_JOURNAL_FILE_SOURCE_TYPE source_type = SDJF_ALL;
size_t filters = 0;
size_t sampling = SYSTEMD_JOURNAL_DEFAULT_ITEMS_SAMPLING;
@@ -1753,9 +1687,6 @@ void function_systemd_journal(const char *transaction, char *function, int timeo
else if(strcmp(keyword, JOURNAL_PARAMETER_INFO) == 0) {
info = true;
}
- else if(strcmp(keyword, JOURNAL_PARAMETER_PROGRESS) == 0) {
- progress = true;
- }
else if(strncmp(keyword, JOURNAL_PARAMETER_DELTA ":", sizeof(JOURNAL_PARAMETER_DELTA ":") - 1) == 0) {
char *v = &keyword[sizeof(JOURNAL_PARAMETER_DELTA ":") - 1];
@@ -1791,12 +1722,6 @@ void function_systemd_journal(const char *transaction, char *function, int timeo
else
slice = true;
}
- else if(strncmp(keyword, JOURNAL_PARAMETER_ID ":", sizeof(JOURNAL_PARAMETER_ID ":") - 1) == 0) {
- char *id = &keyword[sizeof(JOURNAL_PARAMETER_ID ":") - 1];
-
- if(*id)
- progress_id = id;
- }
else if(strncmp(keyword, JOURNAL_PARAMETER_SOURCE ":", sizeof(JOURNAL_PARAMETER_SOURCE ":") - 1) == 0) {
const char *value = &keyword[sizeof(JOURNAL_PARAMETER_SOURCE ":") - 1];
@@ -1930,15 +1855,7 @@ void function_systemd_journal(const char *transaction, char *function, int timeo
// ------------------------------------------------------------------------
// put this request into the progress db
- if(progress_id && *progress_id) {
- fqs_item = dictionary_set_and_acquire_item(function_query_status_dict, progress_id, &tmp_fqs, sizeof(tmp_fqs));
- fqs = dictionary_acquired_item_value(fqs_item);
- }
- else {
- // no progress id given, proceed without registering our progress in the dictionary
- fqs = &tmp_fqs;
- fqs_item = NULL;
- }
+ fqs = &tmp_fqs;
// ------------------------------------------------------------------------
// validate parameters
@@ -1969,6 +1886,7 @@ void function_systemd_journal(const char *transaction, char *function, int timeo
// ------------------------------------------------------------------------
// set query time-frame, anchors and direction
+ fqs->transaction = transaction;
fqs->after_ut = after_s * USEC_PER_SEC;
fqs->before_ut = (before_s * USEC_PER_SEC) + USEC_PER_SEC - 1;
fqs->if_modified_since = if_modified_since;
@@ -2045,11 +1963,9 @@ void function_systemd_journal(const char *transaction, char *function, int timeo
buffer_json_member_add_boolean(wb, JOURNAL_PARAMETER_INFO, false);
buffer_json_member_add_boolean(wb, JOURNAL_PARAMETER_SLICE, fqs->slice);
buffer_json_member_add_boolean(wb, JOURNAL_PARAMETER_DATA_ONLY, fqs->data_only);
- buffer_json_member_add_boolean(wb, JOURNAL_PARAMETER_PROGRESS, false);
buffer_json_member_add_boolean(wb, JOURNAL_PARAMETER_DELTA, fqs->delta);
buffer_json_member_add_boolean(wb, JOURNAL_PARAMETER_TAIL, fqs->tail);
buffer_json_member_add_uint64(wb, JOURNAL_PARAMETER_SAMPLING, fqs->sampling);
- buffer_json_member_add_string(wb, JOURNAL_PARAMETER_ID, progress_id);
buffer_json_member_add_uint64(wb, "source_type", fqs->source_type);
buffer_json_member_add_uint64(wb, JOURNAL_PARAMETER_AFTER, fqs->after_ut / USEC_PER_SEC);
buffer_json_member_add_uint64(wb, JOURNAL_PARAMETER_BEFORE, fqs->before_ut / USEC_PER_SEC);
@@ -2098,11 +2014,6 @@ void function_systemd_journal(const char *transaction, char *function, int timeo
goto output;
}
- if(progress) {
- function_systemd_journal_progress(wb, transaction, progress_id);
- goto cleanup;
- }
-
response = netdata_systemd_journal_query(wb, facets, fqs);
// ------------------------------------------------------------------------
@@ -2124,16 +2035,4 @@ cleanup:
simple_pattern_free(sources);
facets_destroy(facets);
buffer_free(wb);
-
- if(fqs_item) {
- dictionary_del(function_query_status_dict, dictionary_acquired_item_name(fqs_item));
- dictionary_acquired_item_release(function_query_status_dict, fqs_item);
- dictionary_garbage_collect(function_query_status_dict);
- }
-}
-
-void journal_init_query_status(void) {
- function_query_status_dict = dictionary_create_advanced(
- DICT_OPTION_DONT_OVERWRITE_VALUE | DICT_OPTION_FIXED_SIZE,
- NULL, sizeof(FUNCTION_QUERY_STATUS));
}
diff --git a/collectors/systemd-journal.plugin/systemd-units.c b/src/collectors/systemd-journal.plugin/systemd-units.c
index dac158817..0e096dae1 100644
--- a/collectors/systemd-journal.plugin/systemd-units.c
+++ b/src/collectors/systemd-journal.plugin/systemd-units.c
@@ -1207,19 +1207,19 @@ static void systemd_unit_priority(UnitInfo *u, size_t units) {
u->prio = (prio * units) + u->prio;
}
-#define if_less(current, max, target) ({ \
- typeof(current) _wanted = (current); \
- if((current) < (target)) \
- _wanted = (target) > (max) ? (max) : (target); \
- _wanted; \
-})
-
-#define if_normal(current, max, target) ({ \
- typeof(current) _wanted = (current); \
- if((current) == FACET_ROW_SEVERITY_NORMAL) \
- _wanted = (target) > (max) ? (max) : (target); \
- _wanted; \
-})
+static inline FACET_ROW_SEVERITY if_less(FACET_ROW_SEVERITY current, FACET_ROW_SEVERITY max, FACET_ROW_SEVERITY target) {
+ FACET_ROW_SEVERITY wanted = current;
+ if(current < target)
+ wanted = target > max ? max : target;
+ return wanted;
+}
+
+static inline FACET_ROW_SEVERITY if_normal(FACET_ROW_SEVERITY current, FACET_ROW_SEVERITY max, FACET_ROW_SEVERITY target) {
+ FACET_ROW_SEVERITY wanted = current;
+ if(current == FACET_ROW_SEVERITY_NORMAL)
+ wanted = target > max ? max : target;
+ return wanted;
+}
FACET_ROW_SEVERITY system_unit_severity(UnitInfo *u) {
FACET_ROW_SEVERITY severity, max_severity;
@@ -1596,7 +1596,10 @@ void systemd_units_assign_priority(UnitInfo *base) {
}
}
-void function_systemd_units(const char *transaction, char *function, int timeout, bool *cancelled) {
+void function_systemd_units(const char *transaction, char *function,
+ usec_t *stop_monotonic_ut __maybe_unused, bool *cancelled __maybe_unused,
+ BUFFER *payload __maybe_unused, HTTP_ACCESS access __maybe_unused,
+ const char *source __maybe_unused, void *data __maybe_unused) {
char *words[SYSTEMD_UNITS_MAX_PARAMS] = { NULL };
size_t num_words = quoted_strings_splitter_pluginsd(function, words, SYSTEMD_UNITS_MAX_PARAMS);
for(int i = 1; i < SYSTEMD_UNITS_MAX_PARAMS ;i++) {
@@ -1829,7 +1832,7 @@ void function_systemd_units(const char *transaction, char *function, int timeout
case SD_BUS_TYPE_UINT32:
case SD_BUS_TYPE_INT64:
case SD_BUS_TYPE_UINT64: {
- double m;
+ double m = 0.0;
if(unit_attributes[i].value_type == SD_BUS_TYPE_UINT64)
m = (double)max[i].uint64;
else if(unit_attributes[i].value_type == SD_BUS_TYPE_INT64)
diff --git a/collectors/tc.plugin/README.md b/src/collectors/tc.plugin/README.md
index 2a20ff262..2a20ff262 120000
--- a/collectors/tc.plugin/README.md
+++ b/src/collectors/tc.plugin/README.md
diff --git a/collectors/tc.plugin/integrations/tc_qos_classes.md b/src/collectors/tc.plugin/integrations/tc_qos_classes.md
index 7a6650660..0cbcd4ca9 100644
--- a/collectors/tc.plugin/integrations/tc_qos_classes.md
+++ b/src/collectors/tc.plugin/integrations/tc_qos_classes.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/tc.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/tc.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/tc.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/tc.plugin/metadata.yaml"
sidebar_label: "tc QoS classes"
learn_status: "Published"
-learn_rel_path: "Data Collection/Linux Systems/Network"
+learn_rel_path: "Collecting Metrics/Linux Systems/Network"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -117,7 +117,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/tc.plugin/metadata.yaml b/src/collectors/tc.plugin/metadata.yaml
index f4039a8c5..f4039a8c5 100644
--- a/collectors/tc.plugin/metadata.yaml
+++ b/src/collectors/tc.plugin/metadata.yaml
diff --git a/collectors/tc.plugin/plugin_tc.c b/src/collectors/tc.plugin/plugin_tc.c
index eae70453f..eae70453f 100644
--- a/collectors/tc.plugin/plugin_tc.c
+++ b/src/collectors/tc.plugin/plugin_tc.c
diff --git a/collectors/tc.plugin/tc-qos-helper.sh.in b/src/collectors/tc.plugin/tc-qos-helper.sh.in
index 3298c39a3..d7aad40f4 100755
--- a/collectors/tc.plugin/tc-qos-helper.sh.in
+++ b/src/collectors/tc.plugin/tc-qos-helper.sh.in
@@ -154,7 +154,7 @@ if [ ! -d "${fireqos_run_dir}" ]; then
warning "Although FireQOS is installed on this system as '${fireqos}', I cannot find/read its installation configuration at '${fireqos_exec_dir}/install.config'."
fi
else
- warning "FireQOS is not installed on this system. Use FireQOS to apply traffic QoS and expose the class names to netdata. Check https://github.com/netdata/netdata/tree/master/collectors/tc.plugin#tcplugin"
+ warning "FireQOS is not installed on this system. Use FireQOS to apply traffic QoS and expose the class names to netdata. Check https://github.com/netdata/netdata/tree/master/src/collectors/tc.plugin#tcplugin"
fi
fi
diff --git a/collectors/timex.plugin/README.md b/src/collectors/timex.plugin/README.md
index 89c1bd0d4..89c1bd0d4 120000
--- a/collectors/timex.plugin/README.md
+++ b/src/collectors/timex.plugin/README.md
diff --git a/collectors/timex.plugin/integrations/timex.md b/src/collectors/timex.plugin/integrations/timex.md
index 754b2368c..ce0e0ed2c 100644
--- a/collectors/timex.plugin/integrations/timex.md
+++ b/src/collectors/timex.plugin/integrations/timex.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/timex.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/timex.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/timex.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/timex.plugin/metadata.yaml"
sidebar_label: "Timex"
learn_status: "Published"
-learn_rel_path: "Data Collection/System Clock and NTP"
+learn_rel_path: "Collecting Metrics/System Clock and NTP"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -76,7 +76,7 @@ The following alerts are available:
| Alert name | On metric | Description |
|:------------|:----------|:------------|
-| [ system_clock_sync_state ](https://github.com/netdata/netdata/blob/master/health/health.d/timex.conf) | system.clock_sync_state | when set to 0, the system kernel believes the system clock is not properly synchronized to a reliable server |
+| [ system_clock_sync_state ](https://github.com/netdata/netdata/blob/master/src/health/health.d/timex.conf) | system.clock_sync_state | when set to 0, the system kernel believes the system clock is not properly synchronized to a reliable server |
## Setup
@@ -103,7 +103,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/timex.plugin/plugin_timex.c b/src/collectors/timex.plugin/plugin_timex.c
index 025b699a1..025b699a1 100644
--- a/collectors/timex.plugin/plugin_timex.c
+++ b/src/collectors/timex.plugin/plugin_timex.c
diff --git a/collectors/xenstat.plugin/README.md b/src/collectors/xenstat.plugin/README.md
index 32fe4d213..32fe4d213 120000
--- a/collectors/xenstat.plugin/README.md
+++ b/src/collectors/xenstat.plugin/README.md
diff --git a/collectors/xenstat.plugin/integrations/xen_xcp-ng.md b/src/collectors/xenstat.plugin/integrations/xen_xcp-ng.md
index 17dc8d785..8659d97ec 100644
--- a/collectors/xenstat.plugin/integrations/xen_xcp-ng.md
+++ b/src/collectors/xenstat.plugin/integrations/xen_xcp-ng.md
@@ -1,9 +1,9 @@
<!--startmeta
-custom_edit_url: "https://github.com/netdata/netdata/edit/master/collectors/xenstat.plugin/README.md"
-meta_yaml: "https://github.com/netdata/netdata/edit/master/collectors/xenstat.plugin/metadata.yaml"
+custom_edit_url: "https://github.com/netdata/netdata/edit/master/src/collectors/xenstat.plugin/README.md"
+meta_yaml: "https://github.com/netdata/netdata/edit/master/src/collectors/xenstat.plugin/metadata.yaml"
sidebar_label: "Xen XCP-ng"
learn_status: "Published"
-learn_rel_path: "Data Collection/Containers and VMs"
+learn_rel_path: "Collecting Metrics/Containers and VMs"
most_popular: False
message: "DO NOT EDIT THIS FILE DIRECTLY, IT IS GENERATED BY THE COLLECTOR'S metadata.yaml FILE"
endmeta-->
@@ -152,7 +152,7 @@ The file format is a modified INI syntax. The general structure is:
option3 = some third value
```
You can edit the configuration file using the `edit-config` script from the
-Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/configure/nodes.md#the-netdata-config-directory).
+Netdata [config directory](https://github.com/netdata/netdata/blob/master/docs/netdata-agent/configuration.md#the-netdata-config-directory).
```bash
cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
diff --git a/collectors/xenstat.plugin/metadata.yaml b/src/collectors/xenstat.plugin/metadata.yaml
index e5527dbb1..e5527dbb1 100644
--- a/collectors/xenstat.plugin/metadata.yaml
+++ b/src/collectors/xenstat.plugin/metadata.yaml
diff --git a/collectors/xenstat.plugin/xenstat_plugin.c b/src/collectors/xenstat.plugin/xenstat_plugin.c
index 319396d43..3756e112c 100644
--- a/collectors/xenstat.plugin/xenstat_plugin.c
+++ b/src/collectors/xenstat.plugin/xenstat_plugin.c
@@ -975,7 +975,7 @@ int main(int argc, char **argv) {
" --help print this message and exit\n"
"\n"
" For more information:\n"
- " https://github.com/netdata/netdata/tree/master/collectors/xenstat.plugin\n"
+ " https://github.com/netdata/netdata/tree/master/src/collectors/xenstat.plugin\n"
"\n"
, VERSION
, netdata_update_every
@@ -1032,10 +1032,7 @@ int main(int argc, char **argv) {
if(unlikely(netdata_exit)) break;
if(unlikely(debug && iteration))
- fprintf(stderr, "xenstat.plugin: iteration %zu, dt %llu usec\n"
- , iteration
- , dt
- );
+ fprintf(stderr, "xenstat.plugin: iteration %zu, dt %lu usec\n", iteration, dt);
if(likely(xhandle)) {
if(unlikely(debug)) fprintf(stderr, "xenstat.plugin: calling xenstat_collect()\n");