Deploying and Scaling AI Applications with the NVIDIA TensorRT Inference Server on Kubernetes

Similar Tracks
Scaling AI Workloads with Kubernetes: Sharing GPU Resources Across Multiple Containers - Jack Ong
The Linux Foundation
AWS re:Invent 2022 - Deep learning on AWS with NVIDIA: From training to deployment (PRT219)
AWS Events
NVIDIA AI Tech Workshop at NeurIPS Expo 2018 - Session 3: Inference and Quantization
NVIDIA Developer
Everything you Need to Know about using GPUs with Kubernetes - Rohit Agarwal, Google
CNCF [Cloud Native Computing Foundation]
Use Nvidia’s DeepStream and Transfer Learning Toolkit to Deploy Streaming Analytics at Scale
NVIDIA Developer
Unlocking the Full Potential of GPUs for AI Workloads on Kubernetes - Kevin Klues, NVIDIA
CNCF [Cloud Native Computing Foundation]