Similar Tracks
Mastering LLM Inference Optimization From Theory to Cost Effective Deployment: Mark Moyou
AI Engineer
vLLM: Easy, Fast, and Cheap LLM Serving for Everyone - Woosuk Kwon & Xiaoxuan Liu, UC Berkeley
PyTorch
Exploring the Latency/Throughput & Cost Space for LLM Inference // Timothée Lacroix // CTO Mistral
MLOps.community