Similar Tracks
MedAI #54: FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness | Tri Dao
Stanford MedAI
Lecture 1: Swin Transformer from Scratch in PyTorch - Hierarchic Structure and Shifted Windows Ideas
AI HMP
Attention is all you need (Transformer) - Model explanation (including math), Inference and Training
Umar Jamil