QLoRA—How to Fine-tune an LLM on a Single GPU (w/ Python Code) Share: Download MP3 Similar Tracks How to Improve LLMs with RAG (Overview + Python Code) Shaw Talebi Local LLM Fine-tuning on Mac (M1 16GB) Shaw Talebi LoRA explained (and a bit about precision and quantization) DeepFindr Text Embeddings, Classification, and Semantic Search (w/ Python Code) Shaw Talebi Fine-tuning Large Language Models (LLMs) | w/ Example Code Shaw Talebi The Hugging Face Transformers Library | Example Code + Chatbot UI with Gradio Shaw Talebi Fine-Tuning Text Embeddings For Domain-specific Search (w/ Python) Shaw Talebi Fine-tuning LLMs on Human Feedback (RLHF + DPO) Shaw Talebi EASIEST Way to Train LLM Train w/ unsloth (2x faster with 70% less GPU memory required) AI Jason Fine-Tuning BERT for Text Classification (w/ Example Code) Shaw Talebi A Practical Introduction to Large Language Models (LLMs) Shaw Talebi Fine Tune DeepSeek R1 | Build a Medical Chatbot DataCamp How to Build an LLM from Scratch | An Overview Shaw Talebi How to Train LLMs to "Think" (o1 & DeepSeek-R1) Shaw Talebi Compressing Large Language Models (LLMs) | w/ Python Code Shaw Talebi How to Evaluate (and Improve) Your LLM Apps Shaw Talebi Quantization vs Pruning vs Distillation: Optimizing NNs for Inference Efficient NLP An Introduction to AI Agents (for 2025) Shaw Talebi