0
0
Fork 0
mirror of https://github.com/netdata/netdata.git synced 2025-04-16 10:31:07 +00:00
netdata_netdata/collectors/python.d.plugin/nvidia_smi
2019-04-30 18:19:52 +03:00
..
Makefile.inc nvidia_smi: init version added () 2018-11-12 23:47:04 +09:00
nvidia_smi.chart.py nvidia_smi: power chart algorithm,mul,div fix () 2019-04-30 18:19:52 +03:00
nvidia_smi.conf python.d.plugin update () 2018-12-07 10:15:03 +01:00
README.md Ga () 2018-12-07 11:30:04 +01:00

nvidia_smi

This module monitors the nvidia-smi cli tool.

Requirements and Notes:

  • You must have the nvidia-smi tool installed and your NVIDIA GPU(s) must support the tool. Mostly the newer high end models used for AI / ML and Crypto or Pro range, read more about nvidia_smi.

  • You must enable this plugin as its disabled by default due to minor performance issues.

  • On some systems when the GPU is idle the nvidia-smi tool unloads and there is added latency again when it is next queried. If you are running GPUs under constant workload this isn't likely to be an issue.

  • Currently the nvidia-smi tool is being queried via cli. Updating the plugin to use the nvidia c/c++ API directly should resolve this issue. See discussion here: https://github.com/netdata/netdata/pull/4357

  • Contributions are welcome.

  • Make sure netdata user can execute /usr/bin/nvidia-smi or wherever your binary is.

  • poll_seconds is how often in seconds the tool is polled for as an integer.

It produces:

  1. Per GPU
  • GPU utilization
  • memory allocation
  • memory utilization
  • fan speed
  • power usage
  • temperature
  • clock speed
  • PCI bandwidth

configuration

Sample:

poll_seconds: 1

analytics