This is the output from
nvidia-smi -q -d CLOCK:
==============NVSMI LOG==============
Timestamp : Mon Dec 4 19:05:48 2017
Driver Version : 384.98
Attached GPUs : 2
GPU 00000000:01:00.0
Clocks
Graphics : 1657 MHz
SM : 1657 MHz
Memory : 3802 MHz
Video : 1480 MHz
Applications Clocks
Graphics : N/A
Memory : N/A
Default Applications Clocks
Graphics : N/A
Memory : N/A
Max Clocks
Graphics : 1974 MHz
SM : 1974 MHz
Memory : 4004 MHz
Video : 1708 MHz
Max Customer Boost Clocks
Graphics : N/A
SM Clock Samples
Duration : 4.36 sec
Number of Samples : 100
Max : 1733 MHz
Min : 1620 MHz
Avg : 1682 MHz
Memory Clock Samples
Duration : 4.36 sec
Number of Samples : 100
Max : 3802 MHz
Min : 3802 MHz
Avg : 3802 MHz
Clock Policy
Auto Boost : N/A
Auto Boost Default : N/A
- Does
Graphics: 1657 MHz stand for CPU Clock?
- Does
Max Clocks / Graphics: 1900 MHz means I can rise it to 1900 without any harm to the GPU?
I've noticed that chaning
GPUGraphicsMemoryOffset changes
Graphics precisely. Changing
GPUMemoryTransferRateOffset affect
Memory but a bit weird, e.g.:
Attribute 'GPUMemoryTransferRateOffset' (ubuntu:0[gpu:0]) assigned value 400.
Will set
Memory to 3999 (4000 is the max value??).
Can someone with more experience provide some explanation?