Originally published at: NVIDIA TensorRT Inference Server Available Now | NVIDIA Technical Blog
The NVIDIA TensorRT inference server GA version is now available for download in a container from the NVIDIA GPU Cloud container registry. Announced at GTC Japan and part of the NVIDIA TensorRT Hyperscale Inference Platform, the TensorRT inference server is a containerized microservice for data center production deployments. As more and more applications leverage AI,…