Roadmap
86 notebooks. 8 stages. Sequential or pick your entry point.
Foundations & Environment
Set up your environment, understand transformers from scratch, and master data preparation fundamentals.
Full Model Fine-Tuning
Fine-tune complete transformer models for classification and generation. Build custom training loops with mixed precision.
Parameter-Efficient Fine-Tuning
Master LoRA, QLoRA, adapters, and prompt tuning — train LLMs with 100x fewer parameters.
Advanced Optimization
FlashAttention, DeepSpeed ZeRO, FSDP, gradient checkpointing, and instruction tuning at scale.
Alignment & Specialized Techniques
RLHF, DPO, Constitutional AI, reward models, and safety evaluation for aligned LLMs.
Custom Kernels & Production
Write CUDA/Triton kernels, master quantization, implement speculative decoding, and deploy with vLLM.
LLM Inference Optimization
Profile, optimize, and deploy LLM inference at scale — from KV cache to quantization to multi-GPU serving.
2024–2025 Techniques
GRPO reasoning models, ORPO/KTO alignment, Unsloth acceleration, SGLang serving, synthetic data pipelines, model merging, and standardized evaluation.