Learn what��s new in the latest releases of NVIDIA CUDA-X AI libraries and the NGC catalog. For more information, see the package release notes. NVIDIA Triton Inference Server (formerly NVIDIA TensorRT Inference Server) simplifies the deployment of AI models at scale in production. It is an open source inference serving software that lets teams deploy trained AI models from any framework��
]]>