Deploying and Scaling AI Applications with the NVIDIA TensorRT Inference Server on Kubernetes

Similar Tracks
Scaling AI Workloads with Kubernetes: Sharing GPU Resources Across Multiple Containers - Jack Ong
The Linux Foundation
Keynote: Accelerating AI Workloads with GPUs in Kubernetes - Kevin Klues & Sanjay Chatterjee
CNCF [Cloud Native Computing Foundation]
GTC 2020: Deep into Triton Inference Server: BERT Practical Deployment on NVIDIA GPU
Bitcoin Standard
AWS re:Invent 2022 - Deep learning on AWS with NVIDIA: From training to deployment (PRT219)
AWS Events