I want to run a HPL test on our 6P40 prodoct.I tried to install the hpl-2.0FERMI_v15 and run this on my 6p40 or 8*v100 platform. The result was just 10%~20% of the system. And the GPU mem usage was just 10% of the GPU have. Do you have some configuration to make the HPL or CUDA to use all of the GPU memory. Or you have a newer version of the HPL-GPU? Thanks,
Wed Jul 17 18:39:58 2019
±----------------------------------------------------------------------------+
| NVIDIA-SMI 418.67 Driver Version: 418.67 CUDA Version: 10.1 |
|-------------------------------±---------------------±---------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
|===============================+======================+======================|
| 0 Tesla P40 On | 00000000:3B:00.0 Off | 0 |
| N/A 35C P0 49W / 250W | 2285MiB / 22919MiB | 0% Default |
±------------------------------±---------------------±---------------------+
| 1 Tesla P40 On | 00000000:60:00.0 Off | 0 |
| N/A 34C P0 50W / 250W | 2285MiB / 22919MiB | 0% Default |
±------------------------------±---------------------±---------------------+
| 2 Tesla P40 On | 00000000:61:00.0 Off | 0 |
| N/A 36C P0 50W / 250W | 2285MiB / 22919MiB | 0% Default |
±------------------------------±---------------------±---------------------+
| 3 Tesla P40 On | 00000000:86:00.0 Off | 0 |
| N/A 43C P0 52W / 250W | 2285MiB / 22919MiB | 0% Default |
±------------------------------±---------------------±---------------------+
| 4 Tesla P40 On | 00000000:DA:00.0 Off | 0 |
| N/A 34C P0 51W / 250W | 2285MiB / 22919MiB | 0% Default |
±------------------------------±---------------------±---------------------+
| 5 Tesla P40 On | 00000000:DB:00.0 Off | 0 |
| N/A 31C P0 49W / 250W | 2285MiB / 22919MiB | 0% Default |
±------------------------------±---------------------±---------------------+
Wed Jul 17 18:41:55 2019
±----------------------------------------------------------------------------+
| NVIDIA-SMI 418.67 Driver Version: 418.67 CUDA Version: 10.1 |
|-------------------------------±---------------------±---------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
|===============================+======================+======================|
| 0 Tesla V100-SXM2… Off | 00000000:1A:00.0 Off | 0 |
| N/A 44C P0 89W / 300W | 2454MiB / 32480MiB | 100% Default |
±------------------------------±---------------------±---------------------+
| 1 Tesla V100-SXM2… Off | 00000000:1B:00.0 Off | 0 |
| N/A 46C P0 93W / 300W | 2454MiB / 32480MiB | 100% Default |
±------------------------------±---------------------±---------------------+
| 2 Tesla V100-SXM2… Off | 00000000:3D:00.0 Off | 0 |
| N/A 46C P0 77W / 300W | 2454MiB / 32480MiB | 100% Default |
±------------------------------±---------------------±---------------------+
| 3 Tesla V100-SXM2… Off | 00000000:3E:00.0 Off | 0 |
| N/A 43C P0 72W / 300W | 2454MiB / 32480MiB | 18% Default |
±------------------------------±---------------------±---------------------+
| 4 Tesla V100-SXM2… Off | 00000000:88:00.0 Off | 0 |
| N/A 46C P0 121W / 300W | 2454MiB / 32480MiB | 100% Default |
±------------------------------±---------------------±---------------------+
| 5 Tesla V100-SXM2… Off | 00000000:89:00.0 Off | 0 |
| N/A 46C P0 78W / 300W | 2454MiB / 32480MiB | 100% Default |
±------------------------------±---------------------±---------------------+
| 6 Tesla V100-SXM2… Off | 00000000:B2:00.0 Off | 0 |
| N/A 45C P0 129W / 300W | 2454MiB / 32480MiB | 91% Default |
±------------------------------±---------------------±---------------------+
| 7 Tesla V100-SXM2… Off | 00000000:B3:00.0 Off | 0 |
| N/A 41C P0 80W / 300W | 2454MiB / 32480MiB | 1% Default |
±------------------------------±---------------------±---------------------+