You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I want to measure three metrics - all of them in percents:
RAM memory usage as: (total - available) / total * 100
where total is MemTotal and available is MemAvailable from /proc/meminfo.
Disk usage of root ("/").
Disk usage of mounted device ("/dev/sda1") <is it possible using labels or PARTUUID?>.
How can I do that? Please help.
My configuration file:
# netdata configuration
#
# You can download the latest version of this file, using:
#
# wget -O /etc/netdata/netdata.conf http://localhost:19999/netdata.conf
# or
# curl -o /etc/netdata/netdata.conf http://localhost:19999/netdata.conf
#
# You can uncomment and change any of the options below.
# The value shown in the commented settings, is the default value.
#
# global netdata configuration
[global]
run as user = netdata
# option 'web files owner' is not used.
web files owner = root
# option 'web files group' is not used.
web files group = root
# glibc malloc arena max for plugins = 1
# glibc malloc arena max for netdata = 1
# libuv worker threads = 16
hostname = GITHUB
# host access prefix =
# enable metric correlations = yes
# metric correlations method = ks2
timezone = Europe/Warsaw
# OOM score = -900
# process scheduling policy = batch
# process nice level = 19
# pthread stack size = 8388608
[db]
update every = 1
mode = dbengine
dbengine page cache with malloc = yes
dbengine page cache size MB = 32
dbengine disk space MB = 512
# dbengine multihost disk space MB = 256
memory deduplication (ksm) = yes
cleanup obsolete charts after secs = 3600
gap when lost iterations above = 1
# enable replication = yes
# seconds to replicate = 86400
# seconds per replication step = 600
cleanup orphan hosts after secs = 3600
storage tiers = 2
# dbengine page fetch timeout secs = 3
# dbengine page fetch retries = 3
# dbengine page descriptors in file mapped memory = no
dbengine tier 1 page cache size MB = 16
dbengine tier 1 multihost disk space MB = 128
dbengine tier 1 update every iterations = 60
dbengine tier 1 backfill = new
dbengine tier 2 page cache size MB = 8
dbengine tier 2 multihost disk space MB = 384
dbengine tier 2 update every iterations = 600
dbengine tier 2 backfill = new
delete obsolete charts files = yes
delete orphan hosts files = yes
enable zero metrics = no
# replication threads = 1
dbengine pages per extent = 64
[directories]
# config = /etc/netdata
# stock config = /usr/lib/netdata/conf.d
# log = /var/log/netdata
# web = /usr/share/netdata/web
# cache = /var/cache/netdata
# lib = /var/lib/netdata
# home = /var/lib/netdata
# lock = /var/lib/netdata/lock
# plugins = "/usr/lib/netdata/plugins.d" "/etc/netdata/custom-plugins.d"
# registry = /var/lib/netdata/registry
# health config = /etc/netdata/health.d
# stock health config = /usr/lib/netdata/conf.d/health.d
[logs]
# debug flags = 0x0000000000000000
# debug = /var/log/netdata/debug.log
# error = /var/log/netdata/error.log
# access = /var/log/netdata/access.log
# health = /var/log/netdata/health.log
# facility = daemon
# errors flood protection period = 1200
# errors to trigger flood protection = 200
[environment variables]
# PATH = /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/sbin:/usr/sbin:/usr/local/bin:/usr/local/sbin
# PYTHONPATH =
# TZ = :/etc/localtime
[host labels]
# name = value
[sqlite]
# auto vacuum = INCREMENTAL
# synchronous = NORMAL
# journal mode = WAL
# temp store = MEMORY
# journal size limit = 16777216
# cache size = -2000
[health]
# silencers file = /var/lib/netdata/health.silencers.json
# enabled = yes
# default repeat warning = never
# default repeat critical = never
# in memory max health log entries = 1000
# script to execute on alarm = /usr/lib/netdata/plugins.d/alarm-notify.sh
# enable stock health configuration = yes
# run at least every seconds = 10
# postpone alarms during hibernation for seconds = 60
# rotate log every lines = 2000
# is ephemeral = no
# has unstable connection = no
[web]
bind to = 0.0.0.0
# ssl key = /etc/netdata/ssl/key.pem
# ssl certificate = /etc/netdata/ssl/cert.pem
# tls version = 1.3
# tls ciphers = none
# ses max window = 15
# des max window = 15
# mode = static-threaded
# listen backlog = 4096
# default port = 19999
# disconnect idle clients after seconds = 60
# timeout for first request = 60
# accept a streaming request every seconds = 0
# respect do not track policy = no
# x-frame-options response header =
# allow connections from = localhost *
# allow connections by dns = heuristic
# allow dashboard from = localhost *
# allow dashboard by dns = heuristic
# allow badges from = *
# allow badges by dns = heuristic
# allow streaming from = *
# allow streaming by dns = heuristic
# allow netdata.conf from = localhost fd* 10.* 192.168.* 172.16.* 172.17.* 172.18.* 172.19.* 172.20.* 172.21.* 172.22.* 172.23.* 172.24.* 172.25.* 172.26.* 172.27.* 172.28.* 172.29.* 172.30.* 172.31.* UNKNOWN
# allow netdata.conf by dns = no
# allow management from = localhost
# allow management by dns = heuristic
# enable gzip compression = yes
# gzip compression strategy = default
# gzip compression level = 3
# web server threads = 4
# web server max sockets = 16384
# custom dashboard_info.js =
[registry]
# enabled = no
# netdata unique id file = /var/lib/netdata/registry/netdata.public.unique.id
# registry db file = /var/lib/netdata/registry/registry.db
# registry log file = /var/lib/netdata/registry/registry-log.db
# registry save db every new entries = 1000000
# registry expire idle persons days = 365
# registry domain =
# registry to announce = https://registry.my-netdata.io
# registry hostname = LIPA-WIREHOLE
# verify browser cookies support = yes
# enable cookies SameSite and Secure = yes
# max URL length = 1024
# max URL name length = 50
# netdata management api key file = /var/lib/netdata/netdata.api.key
# allow from = *
# allow by dns = heuristic
[global statistics]
# update every = 1
[plugins]
timex = no
idlejitter = no
netdata monitoring = no
tc = no
diskspace = no
# proc = yes
cgroups = no
# enable running new plugins = yes
# check for new plugins every = 60
slabinfo = no
apps = no
statsd = no
cups = no
perf = no
nfacct = no
python.d = no
charts.d = no
fping = no
ioping = no
[statsd]
# update every (flushInterval) = 1
# udp messages to process at once = 10
# create private charts for metrics matching = *
# max private charts hard limit = 1000
# private charts history = 3600
# decimal detail = 1000
# disconnect idle tcp clients after seconds = 600
# private charts hidden = no
# histograms and timers percentile (percentThreshold) = 95.00000
# dictionaries max unique dimensions = 200
# add dimension for number of events received = no
# gaps on gauges (deleteGauges) = no
# gaps on counters (deleteCounters) = no
# gaps on meters (deleteMeters) = no
# gaps on sets (deleteSets) = no
# gaps on histograms (deleteHistograms) = no
# gaps on timers (deleteTimers) = no
# gaps on dictionaries (deleteDictionaries) = no
# statsd server max TCP sockets = 16384
# listen backlog = 4096
# default port = 8125
# bind to = udp:localhost tcp:localhost
[plugin:timex]
# update every = 10
# clock synchronization state = yes
# time offset = yes
[plugin:idlejitter]
# loop time in ms = 20
[plugin:apps]
# update every = 1
# command options =
[plugin:cups]
# update every = 1
# command options =
[plugin:perf]
# update every = 1
# command options =
[plugin:nfacct]
# update every = 1
# command options =
[plugin:python.d]
# update every = 1
# command options =
[plugin:charts.d]
# update every = 1
# command options =
[plugin:fping]
# update every = 1
# command options =
[plugin:ioping]
# update every = 1
# command options =
[plugin:tc]
# script to run to get tc values = /usr/lib/netdata/plugins.d/tc-qos-helper.sh
# enable new interfaces detected at runtime = yes
# enable traffic charts for all interfaces = auto
# enable packets charts for all interfaces = auto
# enable dropped charts for all interfaces = auto
# enable tokens charts for all interfaces = no
# enable ctokens charts for all interfaces = no
# enable show all classes and qdiscs for all interfaces = no
# qos for eth0 = yes
# traffic chart for eth0 = auto
# packets chart for eth0 = auto
# dropped packets chart for eth0 = auto
# tokens chart for eth0 = no
# ctokens chart for eth0 = no
# show all classes for eth0 = no
# cleanup unused classes every = 120
[plugin:cgroups]
# update every = 1
# check for new cgroups every = 10
# use unified cgroups = auto
# containers priority = 40000
# enable cpuacct stat (total CPU) = auto
# enable cpuacct usage (per core CPU) = auto
# enable cpuacct cpu throttling = yes
# enable cpuacct cpu shares = no
# enable memory = auto
# enable detailed memory = auto
# enable memory limits fail count = auto
# enable swap memory = auto
# enable blkio bandwidth = auto
# enable blkio operations = auto
# enable blkio throttle bandwidth = auto
# enable blkio throttle operations = auto
# enable blkio queued operations = auto
# enable blkio merged operations = auto
# enable cpu pressure = auto
# enable io some pressure = auto
# enable io full pressure = auto
# enable memory some pressure = auto
# enable memory full pressure = auto
# recheck zero blkio every iterations = 10
# recheck zero memory failcnt every iterations = 10
# recheck zero detailed memory every iterations = 10
# enable systemd services = yes
# enable systemd services detailed memory = no
# report used memory = yes
# path to unified cgroups = /sys/fs/cgroup
# max cgroups to allow = 1000
# max cgroups depth to monitor = 0
# enable by default cgroups matching = !*/init.scope !/system.slice/run-*.scope *.scope /machine.slice/*.service */kubepods/pod*/* */kubepods/*/pod*/* */*-kubepods-pod*/* */*-kubepods-*-pod*/* !*kubepods* !*kubelet* !*/vcpu* !*/emulator !*.mount !*.partition !*.service !*.socket !*.slice !*.swap !*.user !/ !/docker !*/libvirt !/lxc !/lxc/*/* !/lxc.monitor* !/lxc.pivot !/lxc.payload !/machine !/qemu !/system !/systemd !/user *
# enable by default cgroups names matching = *
# search for cgroups in subpaths matching = !*/init.scope !*-qemu !*.libvirt-qemu !/init.scope !/system !/systemd !/user !/user.slice !/lxc/*/* !/lxc.monitor !/lxc.payload/*/* !/lxc.payload.* *
# script to get cgroup names = /usr/lib/netdata/plugins.d/cgroup-name.sh
# script to get cgroup network interfaces = /usr/lib/netdata/plugins.d/cgroup-network
# run script to rename cgroups matching = !/ !*.mount !*.socket !*.partition /machine.slice/*.service !*.service !*.slice !*.swap !*.user !init.scope !*.scope/vcpu* !*.scope/emulator *.scope *docker* *lxc* *qemu* */kubepods/pod*/* */kubepods/*/pod*/* */*-kubepods-pod*/* */*-kubepods-*-pod*/* !*kubepods* !*kubelet* *.libvirt-qemu *
# cgroups to match as systemd services = !/system.slice/*/*.service /system.slice/*.service
[plugin:proc]
# /proc/net/dev = yes
/proc/pagetypeinfo = no
# /proc/stat = yes
/proc/uptime = no
/proc/loadavg = no
/proc/sys/kernel/random/entropy_avail = no
/proc/pressure = no
/proc/interrupts = no
/proc/softirqs = no
/proc/vmstat = no
/proc/meminfo = no
/sys/kernel/mm/ksm = no
/sys/block/zram = no
/sys/devices/system/edac/mc = no
/sys/devices/system/node = no
/proc/net/wireless = no
/proc/net/sockstat = no
/proc/net/sockstat6 = no
/proc/net/netstat = no
/proc/net/sctp/snmp = no
/proc/net/softnet_stat = no
/proc/net/ip_vs/stats = no
/sys/class/infiniband = no
/proc/net/stat/conntrack = no
/proc/net/stat/synproxy = no
/proc/diskstats = no
/proc/mdstat = no
/proc/net/rpc/nfsd = no
/proc/net/rpc/nfs = no
/proc/spl/kstat/zfs/arcstats = no
/proc/spl/kstat/zfs/pool/state = no
/sys/fs/btrfs = no
ipc = no
/sys/class/power_supply = no
[plugin:proc:diskspace]
# remove charts of unmounted disks = yes
# update every = 1
# check for new mount points every = 15
# exclude space metrics on paths = /proc/* /sys/* /var/run/user/* /run/user/* /snap/* /var/lib/docker/*
# exclude space metrics on filesystems = *gvfs *gluster* *s3fs *ipfs *davfs2 *httpfs *sshfs *gdfs *moosefs fusectl autofs
# space usage for all disks = auto
# inodes usage for all disks = auto
[plugin:proc:/proc/stat]
# cpu utilization = yes
per cpu core utilization = no
cpu interrupts = no
context switches = no
processes started = no
processes running = no
keep per core files open = no
keep cpuidle files open = no
core_throttle_count = no
package_throttle_count = no
cpu frequency = no
cpu idle states = no
# core_throttle_count filename to monitor = /sys/devices/system/cpu/%s/thermal_throttle/core_throttle_count
# package_throttle_count filename to monitor = /sys/devices/system/cpu/%s/thermal_throttle/package_throttle_count
# scaling_cur_freq filename to monitor = /sys/devices/system/cpu/%s/cpufreq/scaling_cur_freq
# time_in_state filename to monitor = /sys/devices/system/cpu/%s/cpufreq/stats/time_in_state
# schedstat filename to monitor = /proc/schedstat
# cpuidle name filename to monitor = /sys/devices/system/cpu/cpu%zu/cpuidle/state%zu/name
# cpuidle time filename to monitor = /sys/devices/system/cpu/cpu%zu/cpuidle/state%zu/time
# filename to monitor = /proc/stat
[plugin:proc:/proc/uptime]
# filename to monitor = /proc/uptime
[plugin:proc:/proc/loadavg]
# filename to monitor = /proc/loadavg
# enable load average = yes
# enable total processes = yes
[plugin:proc:/proc/sys/kernel/random/entropy_avail]
# filename to monitor = /proc/sys/kernel/random/entropy_avail
[plugin:proc:/proc/pressure]
# base path of pressure metrics = /proc/pressure
# enable cpu some pressure = yes
# enable cpu full pressure = yes
# enable memory some pressure = yes
# enable memory full pressure = yes
# enable io some pressure = yes
# enable io full pressure = yes
[plugin:proc:/proc/interrupts]
# interrupts per core = auto
# filename to monitor = /proc/interrupts
[plugin:proc:/proc/softirqs]
# interrupts per core = auto
# filename to monitor = /proc/softirqs
[plugin:proc:/proc/vmstat]
# filename to monitor = /proc/vmstat
# swap i/o = auto
# disk i/o = yes
# memory page faults = yes
# out of memory kills = yes
# system-wide numa metric summary = auto
[plugin:proc:/sys/devices/system/node]
# directory to monitor = /sys/devices/system/node
[plugin:proc:/proc/meminfo]
# system ram = yes
# system swap = auto
# hardware corrupted ECC = auto
# committed memory = yes
# writeback memory = yes
# kernel memory = yes
# slab memory = yes
# hugepages = auto
# transparent hugepages = auto
# filename to monitor = /proc/meminfo
[plugin:proc:/sys/kernel/mm/ksm]
# /sys/kernel/mm/ksm/pages_shared = /sys/kernel/mm/ksm/pages_shared
# /sys/kernel/mm/ksm/pages_sharing = /sys/kernel/mm/ksm/pages_sharing
# /sys/kernel/mm/ksm/pages_unshared = /sys/kernel/mm/ksm/pages_unshared
# /sys/kernel/mm/ksm/pages_volatile = /sys/kernel/mm/ksm/pages_volatile
[plugin:proc:/sys/devices/system/edac/mc]
# directory to monitor = /sys/devices/system/edac/mc
[plugin:proc:/proc/net/wireless]
# filename to monitor = /proc/net/wireless
# status for all interfaces = auto
# quality for all interfaces = auto
# discarded packets for all interfaces = auto
# missed beacon for all interface = auto
[plugin:proc:/proc/net/sockstat]
# ipv4 sockets = auto
# ipv4 TCP sockets = auto
# ipv4 TCP memory = auto
# ipv4 UDP sockets = auto
# ipv4 UDP memory = auto
# ipv4 UDPLITE sockets = auto
# ipv4 RAW sockets = auto
# ipv4 FRAG sockets = auto
# ipv4 FRAG memory = auto
# update constants every = 60
# filename to monitor = /proc/net/sockstat
[plugin:proc:/proc/net/sockstat6]
# ipv6 TCP sockets = auto
# ipv6 UDP sockets = auto
# ipv6 UDPLITE sockets = auto
# ipv6 RAW sockets = auto
# ipv6 FRAG sockets = auto
# filename to monitor = /proc/net/sockstat6
[plugin:proc:/proc/net/netstat]
# bandwidth = auto
# input errors = auto
# multicast bandwidth = auto
# broadcast bandwidth = auto
# multicast packets = auto
# broadcast packets = auto
# ECN packets = auto
# TCP reorders = auto
# TCP SYN cookies = auto
# TCP out-of-order queue = auto
# TCP connection aborts = auto
# TCP memory pressures = auto
# TCP SYN queue = auto
# TCP accept queue = auto
# filename to monitor = /proc/net/netstat
[plugin:proc:/proc/net/snmp]
# ipv4 packets = auto
# ipv4 fragments sent = auto
# ipv4 fragments assembly = auto
# ipv4 errors = auto
# ipv4 TCP connections = auto
# ipv4 TCP packets = auto
# ipv4 TCP errors = auto
# ipv4 TCP opens = auto
# ipv4 TCP handshake issues = auto
# ipv4 UDP packets = auto
# ipv4 UDP errors = auto
# ipv4 ICMP packets = auto
# ipv4 ICMP messages = auto
# ipv4 UDPLite packets = auto
# filename to monitor = /proc/net/snmp
[plugin:proc:/proc/net/snmp6]
# ipv6 packets = auto
# ipv6 fragments sent = auto
# ipv6 fragments assembly = auto
# ipv6 errors = auto
# ipv6 UDP packets = auto
# ipv6 UDP errors = auto
# ipv6 UDPlite packets = auto
# ipv6 UDPlite errors = auto
# bandwidth = auto
# multicast bandwidth = auto
# broadcast bandwidth = auto
# multicast packets = auto
# icmp = auto
# icmp redirects = auto
# icmp errors = auto
# icmp echos = auto
# icmp group membership = auto
# icmp router = auto
# icmp neighbor = auto
# icmp mldv2 = auto
# icmp types = auto
# ect = auto
# filename to monitor = /proc/net/snmp6
[plugin:proc:/proc/net/sctp/snmp]
# established associations = auto
# association transitions = auto
# fragmentation = auto
# packets = auto
# packet errors = auto
# chunk types = auto
# filename to monitor = /proc/net/sctp/snmp
[plugin:proc:/proc/net/softnet_stat]
# softnet_stat per core = yes
# filename to monitor = /proc/net/softnet_stat
[plugin:proc:/proc/net/ip_vs_stats]
# IPVS bandwidth = yes
# IPVS connections = yes
# IPVS packets = yes
# filename to monitor = /proc/net/ip_vs_stats
[plugin:proc:/sys/class/infiniband]
# dirname to monitor = /sys/class/infiniband
# bandwidth counters = yes
# packets counters = yes
# errors counters = yes
# hardware packets counters = auto
# hardware errors counters = auto
# monitor only active ports = auto
# disable by default interfaces matching =
# refresh ports state every seconds = 30
[plugin:proc:/proc/net/stat/nf_conntrack]
# filename to monitor = /proc/net/stat/nf_conntrack
# netfilter new connections = no
# netfilter connection changes = no
# netfilter connection expectations = no
# netfilter connection searches = no
# netfilter errors = no
# netfilter connections = yes
[plugin:proc:/proc/sys/net/netfilter/nf_conntrack_max]
# filename to monitor = /proc/sys/net/netfilter/nf_conntrack_max
# read every seconds = 10
[plugin:proc:/proc/sys/net/netfilter/nf_conntrack_count]
# filename to monitor = /proc/sys/net/netfilter/nf_conntrack_count
[plugin:proc:/proc/net/stat/synproxy]
# SYNPROXY cookies = auto
# SYNPROXY SYN received = auto
# SYNPROXY connections reopened = auto
# filename to monitor = /proc/net/stat/synproxy
[plugin:proc:/proc/diskstats]
# enable new disks detected at runtime = yes
# performance metrics for physical disks = auto
# performance metrics for virtual disks = auto
# performance metrics for partitions = no
# bandwidth for all disks = auto
# operations for all disks = auto
# merged operations for all disks = auto
# i/o time for all disks = auto
# queued operations for all disks = auto
# utilization percentage for all disks = auto
# extended operations for all disks = auto
# backlog for all disks = auto
# bcache for all disks = auto
# bcache priority stats update every = 0
# remove charts of removed disks = yes
# path to get block device = /sys/block/%s
# path to get block device bcache = /sys/block/%s/bcache
# path to get virtual block device = /sys/devices/virtual/block/%s
# path to get block device infos = /sys/dev/block/%lu:%lu/%s
# path to device mapper = /dev/mapper
# path to /dev/disk/by-label = /dev/disk/by-label
# path to /dev/disk/by-id = /dev/disk/by-id
# path to /dev/vx/dsk = /dev/vx/dsk
# name disks by id = no
# preferred disk ids = *
# exclude disks = loop* ram*
# filename to monitor = /proc/diskstats
# performance metrics for disks with major 179 = yes
# performance metrics for disks with major 8 = yes
[plugin:proc:/proc/diskstats:ram0]
# enable = no
[plugin:proc:/proc/diskstats:ram1]
# enable = no
[plugin:proc:/proc/diskstats:ram2]
# enable = no
[plugin:proc:/proc/diskstats:ram3]
# enable = no
[plugin:proc:/proc/diskstats:ram4]
# enable = no
[plugin:proc:/proc/diskstats:ram5]
# enable = no
[plugin:proc:/proc/diskstats:ram6]
# enable = no
[plugin:proc:/proc/diskstats:ram7]
# enable = no
[plugin:proc:/proc/diskstats:ram8]
# enable = no
[plugin:proc:/proc/diskstats:ram9]
# enable = no
[plugin:proc:/proc/diskstats:ram10]
# enable = no
[plugin:proc:/proc/diskstats:ram11]
# enable = no
[plugin:proc:/proc/diskstats:ram12]
# enable = no
[plugin:proc:/proc/diskstats:ram13]
# enable = no
[plugin:proc:/proc/diskstats:ram14]
# enable = no
[plugin:proc:/proc/diskstats:ram15]
# enable = no
[plugin:proc:/proc/diskstats:loop0]
# enable = no
[plugin:proc:/proc/diskstats:loop1]
# enable = no
[plugin:proc:/proc/diskstats:loop2]
# enable = no
[plugin:proc:/proc/diskstats:loop3]
# enable = no
[plugin:proc:/proc/diskstats:loop4]
# enable = no
[plugin:proc:/proc/diskstats:loop5]
# enable = no
[plugin:proc:/proc/diskstats:loop6]
# enable = no
[plugin:proc:/proc/diskstats:loop7]
# enable = no
[plugin:proc:/proc/diskstats:mmcblk0]
# enable = yes
# enable performance metrics = yes
# bandwidth = auto
# operations = auto
# merged operations = auto
# i/o time = auto
# queued operations = auto
# utilization percentage = auto
# extended operations = auto
# backlog = auto
[plugin:proc:/proc/diskstats:bootfs]
# enable = yes
# enable performance metrics = no
# bandwidth = no
# operations = no
# merged operations = no
# i/o time = no
# queued operations = no
# utilization percentage = no
# extended operations = no
# backlog = no
[plugin:proc:/proc/diskstats:rootfs]
# enable = yes
# enable performance metrics = no
# bandwidth = no
# operations = no
# merged operations = no
# i/o time = no
# queued operations = no
# utilization percentage = no
# extended operations = no
# backlog = no
[plugin:proc:/proc/diskstats:sda]
# enable = yes
# enable performance metrics = yes
# bandwidth = auto
# operations = auto
# merged operations = auto
# i/o time = auto
# queued operations = auto
# utilization percentage = auto
# extended operations = auto
# backlog = auto
[plugin:proc:/proc/diskstats:sdb]
# enable = yes
# enable performance metrics = yes
# bandwidth = auto
# operations = auto
# merged operations = auto
# i/o time = auto
# queued operations = auto
# utilization percentage = auto
# extended operations = auto
# backlog = auto
[plugin:proc:/proc/diskstats:sdc]
# enable = yes
# enable performance metrics = yes
# bandwidth = auto
# operations = auto
# merged operations = auto
# i/o time = auto
# queued operations = auto
# utilization percentage = auto
# extended operations = auto
# backlog = auto
[plugin:proc:/proc/diskstats:backup]
# enable = yes
# enable performance metrics = no
# bandwidth = no
# operations = no
# merged operations = no
# i/o time = no
# queued operations = no
# utilization percentage = no
# extended operations = no
# backlog = no
[plugin:proc:/proc/diskstats:backup2]
# enable = yes
# enable performance metrics = no
# bandwidth = no
# operations = no
# merged operations = no
# i/o time = no
# queued operations = no
# utilization percentage = no
# extended operations = no
# backlog = no
[plugin:proc:/proc/diskstats:backup3]
# enable = yes
# enable performance metrics = no
# bandwidth = no
# operations = no
# merged operations = no
# i/o time = no
# queued operations = no
# utilization percentage = no
# extended operations = no
# backlog = no
[plugin:proc:/proc/mdstat]
# faulty devices = yes
# nonredundant arrays availability = yes
# mismatch count = auto
# disk stats = yes
# operation status = yes
# make charts obsolete = yes
# filename to monitor = /proc/mdstat
# mismatch_cnt filename to monitor = /sys/block/%s/md/mismatch_cnt
[plugin:proc:/proc/net/rpc/nfsd]
# filename to monitor = /proc/net/rpc/nfsd
# read cache = yes
# file handles = yes
# I/O = yes
# threads = yes
# network = yes
# rpc = yes
# NFS v2 procedures = yes
# NFS v3 procedures = yes
# NFS v4 procedures = yes
# NFS v4 operations = yes
[plugin:proc:/proc/net/rpc/nfs]
# filename to monitor = /proc/net/rpc/nfs
# network = yes
# rpc = yes
# NFS v2 procedures = yes
# NFS v3 procedures = yes
# NFS v4 procedures = yes
[plugin:proc:/proc/spl/kstat/zfs/arcstats]
# filename to monitor = /proc/spl/kstat/zfs/arcstats
[plugin:proc:/proc/spl/kstat/zfs]
# directory to monitor = /proc/spl/kstat/zfs
[plugin:proc:/sys/fs/btrfs]
# path to monitor = /sys/fs/btrfs
# check for btrfs changes every = 60
# physical disks allocation = auto
# data allocation = auto
# metadata allocation = auto
# system allocation = auto
[plugin:proc:ipc]
# message queues = yes
# semaphore totals = yes
# shared memory totals = yes
# msg filename to monitor = /proc/sysvipc/msg
# shm filename to monitor = /proc/sysvipc/shm
# max dimensions in memory allowed = 50
[plugin:proc:/sys/class/power_supply]
# battery capacity = yes
# battery charge = no
# battery energy = no
# power supply voltage = no
# keep files open = auto
# directory to monitor = /sys/class/power_supply
[plugin:proc:/proc/net/dev]
# filename to monitor = /proc/net/dev
# path to get virtual interfaces = /sys/devices/virtual/net/%s
# path to get net device speed = /sys/class/net/%s/speed
# path to get net device duplex = /sys/class/net/%s/duplex
# path to get net device operstate = /sys/class/net/%s/operstate
# path to get net device carrier = /sys/class/net/%s/carrier
# path to get net device mtu = /sys/class/net/%s/mtu
# enable new interfaces detected at runtime = auto
# bandwidth for all interfaces = auto
# packets for all interfaces = auto
# errors for all interfaces = auto
# drops for all interfaces = auto
# fifo for all interfaces = auto
# compressed packets for all interfaces = auto
# frames, collisions, carrier counters for all interfaces = auto
# speed for all interfaces = auto
# duplex for all interfaces = auto
# operstate for all interfaces = auto
# carrier for all interfaces = auto
# mtu for all interfaces = auto
# disable by default interfaces matching = lo fireqos* *-ifb fwpr* fwbr* fwln*
[plugin:proc:/proc/net/dev:lo]
# enabled = no
# virtual = yes
[plugin:proc:/proc/net/dev:eth0]
# enabled = yes
# virtual = no
# bandwidth = auto
packets = no
errors = no
drops = no
fifo = no
compressed = no
events = no
speed = no
duplex = no
operstate = no
carrier = no
mtu = no
[plugin:proc:/proc/net/dev:wlan0]
# enabled = yes
# virtual = no
# bandwidth = auto
packets = no
errors = no
drops = no
fifo = no
compressed = no
events = no
speed = no
duplex = no
operstate = no
carrier = no
mtu = no
[plugin:proc:/proc/net/dev:wg0]
# enabled = yes
# virtual = yes
# bandwidth = auto
packets = no
errors = no
drops = no
fifo = no
compressed = no
events = no
speed = no
duplex = no
operstate = no
carrier = no
mtu = no
[plugin:proc:diskspace:/]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/dev]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/dev/shm]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/run]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/run/lock]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/run/user/1000]
# space usage = no
# inodes usage = no
[plugin:proc:diskspace:/sys/kernel/security]
# space usage = no
# inodes usage = no
[plugin:proc:diskspace:/sys/fs/cgroup]
# space usage = no
# inodes usage = no
[plugin:proc:diskspace:/sys/fs/pstore]
# space usage = no
# inodes usage = no
[plugin:proc:diskspace:/sys/fs/bpf]
# space usage = no
# inodes usage = no
[plugin:proc:diskspace:/sys/kernel/tracing]
# space usage = no
# inodes usage = no
[plugin:proc:diskspace:/sys/kernel/config]
# space usage = no
# inodes usage = no
[plugin:proc:diskspace:/proc/fs/nfsd]
# space usage = no
# inodes usage = no
[plugin:proc:diskspace:/boot/firmware]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/media/BACKUP]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/media/BACKUP2]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/media/BACKUP3]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/mnt/LIPA-NAS]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/run/credentials]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/run/netdata]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/run/systemd/incoming]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/run/user]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/tmp]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/var/cache/netdata]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/var/lib/netdata]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/var/log]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/var/spool]
# space usage = auto
# inodes usage = auto
[plugin:proc:diskspace:/var/tmp]
# space usage = auto
# inodes usage = auto
```
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi!
I'm using Debian 12.
I installed NetData directly from APT-GET.
I want to measure three metrics - all of them in percents:
(total - available) / total * 100where total is MemTotal and available is MemAvailable from /proc/meminfo.
How can I do that? Please help.
My configuration file:
Beta Was this translation helpful? Give feedback.
All reactions