thank you for the quick answer. Unfortunately, it doesn’t answer my question.
I am sorry I haven’t been clear enough, so I’ll reformulate.
I want to use nvidia-persistenced (not nvidia-smi) to set up my persistence mode.
If I try to do that from bash, it complains that there’s already a nvidia-persistenced daemon running with the persistence mode switched off. Indeed, if check with ps I get the following:
Now, I suppose that this daemon is started by Nvidia at boot time. I want the daemon to be started in the same way, but with the persistence mode on, so my questions are:
Is there a Nvidia configuration file/script that I must edit to control the startup of the daemon?
Uncompress the archive
*Edit the nvidia-persistenced.conf.template in the subdirectory corresponding to your startup service (can be upstart, systemd or sysv)
add “–persistence-mode --verbose” to the line where the service is started
run install.sh to kill the currently running daemon and install the config with the new parameters
No instructions there on how to do it; just a description of what it’s supposed to do. Running nvidia-persistenced as superuser spits an error (see above).
My question is nvidia.persistenced brings totally different speed effects on the training oxford_flowers102.
For instance, I use two machines to train 100 epochs AlexNet for TensorFlow.
Machine One: Nvidia RTX 2070 Super
1. Power & GPU Memory:
Initial Power: 3W/215W
Power in Training: 187W/215W
Initial GPU Memory: 300MiB
GPU Memory during Training: 5500MiB
Effect: 26 minutes for completing the training.
It increase form 3W to 187w in the power consumption. In addition, the GPU ďĽemory consumption increases quite fast. Therefore, its training speed is much more fast.
2. Processes:
Beside of GID, it shows GI ID and CI ID with the command of nvidia-smi.
Machine Two: Nvidia RTX 2060
1. Power & GPU Memory
Initial Power: 9W/160W
Power in Training: 16W/160W
Initial GPU Memory: 300MiB
GPU Memory during Training: 500MiB
Its power constantly fixes on 16W and GPU memory grows quite slow. And its training speed is about 4 hours, much more slower than RTX 2070.
2. Processes:
It shows GID but no GI ID and CI ID.
My question is what is GI ID and CI ID. Do they influence the training speed?
Note:
I enable GPU Fan Settings from 45% to 74%. But is has no effect on the training speed except the a little big noise.