# Nvidia GPU monitoring with Netdata Monitors performance metrics (memory usage, fan speed, pcie bandwidth utilization, temperature, etc.) using `nvidia-smi` cli tool. > **Warning**: this collector does not work when the Netdata Agent is [running in a container](https://learn.netdata.cloud/docs/agent/packaging/docker). ## Requirements and Notes - You must have the `nvidia-smi` tool installed and your NVIDIA GPU(s) must support the tool. Mostly the newer high end models used for AI / ML and Crypto or Pro range, read more about [nvidia_smi](https://developer.nvidia.com/nvidia-system-management-interface). - You must enable this plugin, as its disabled by default due to minor performance issues: ```bash cd /etc/netdata # Replace this path with your Netdata config directory, if different sudo ./edit-config python.d.conf ``` Remove the '#' before nvidia_smi so it reads: `nvidia_smi: yes`. - On some systems when the GPU is idle the `nvidia-smi` tool unloads and there is added latency again when it is next queried. If you are running GPUs under constant workload this isn't likely to be an issue. - Currently the `nvidia-smi` tool is being queried via cli. Updating the plugin to use the nvidia c/c++ API directly should resolve this issue. See discussion here: - Contributions are welcome. - Make sure `netdata` user can execute `/usr/bin/nvidia-smi` or wherever your binary is. - If `nvidia-smi` process [is not killed after netdata restart](https://github.com/netdata/netdata/issues/7143) you need to off `loop_mode`. - `poll_seconds` is how often in seconds the tool is polled for as an integer. ## Charts It produces the following charts: - PCI Express Bandwidth Utilization in `KiB/s` - Fan Speed in `percentage` - GPU Utilization in `percentage` - Memory Bandwidth Utilization in `percentage` - Encoder/Decoder Utilization in `percentage` - Memory Usage in `MiB` - Temperature in `celsius` - Clock Frequencies in `MHz` - Power Utilization in `Watts` - Memory Used by Each Process in `MiB` - Memory Used by Each User in `MiB` - Number of User on GPU in `num` ## Configuration Edit the `python.d/nvidia_smi.conf` configuration file using `edit-config` from the Netdata [config directory](/docs/configure/nodes.md), which is typically at `/etc/netdata`. ```bash cd /etc/netdata # Replace this path with your Netdata config directory, if different sudo ./edit-config python.d/nvidia_smi.conf ``` Sample: ```yaml loop_mode : yes poll_seconds : 1 exclude_zero_memory_users : yes ```