I have a A40 NVIDIA GPU on a server, installing OS and Kubernetes over that I created different Docker images and run different Cuda version and different tensorflow how to implement please give detailed solutions.
For just docker:
- install the latest driver for your GPU in the base machine, not in any container.
- make sure you are using a recent version of docker, 19.03ce or newer
- use the nvidia container toolkit
- Install the CUDA toolkit version of choice in the container (do not install NVIDIA GPU driver in any docker container)
- When launching containers, specify
--gps=...
switch (e.g.--gpus=all
) - Profit!
To setup Kubernetes, I would use the cloud native stack.