Stability Issues with GPU Inference on Older GPUs (e.g., 1080Ti)

I’ve no experience with Cuda programs where the GPU is underutilised to this degree.
While it seems GPUs later than the 10XX have clock locking capabilties, you don’t seem to have that option.

This issue has come up from time to time - here for example, but like you have found, it appears that a minimum level of “busy-ness” is required.