Deploying and Scaling AI Applications with the NVIDIA TensorRT Inference Server on Kubernetes

Similar Tracks
GTC 2020: Deep into Triton Inference Server: BERT Practical Deployment on NVIDIA GPU
Bitcoin Standard
Keynote: Accelerating AI Workloads with GPUs in Kubernetes - Kevin Klues & Sanjay Chatterjee
CNCF [Cloud Native Computing Foundation]
Scaling AI Workloads with Kubernetes: Sharing GPU Resources Across Multiple Containers - Jack Ong
The Linux Foundation