This condensed syllabus progresses from basics to advanced AI research, following Karpathy's playlist with minor flow tweaks. Format: 2-hour watch and ask sessions.

Quick Refreshers on Core Concepts

Block 1: Neural Network Fundamentals

  1. The spelled-out intro to neural networks and backpropagation: building micrograd (2:25:52)Key: Auto-diff, loss—interpret decisions. Additional: Linear algebra; calculus basics.

Block 2: Language Modeling Basics

  1. The spelled-out intro to language modeling: building makemore (1:57:45)Key: Bigrams, sampling—model identities.Additional: Probability; loss functions.

Block 3: Scaling to Multi-Layer Networks

  1. Building makemore Part 2: MLP (1:15:40) – Non-linearities.
  2. Building makemore Part 3: Activations & Gradients, BatchNorm (1:55:58) – Stabilize training.
  3. Building makemore Part 4: Becoming a Backprop Ninja (1:55:24) – Chain rule.Additional: Optimization (SGD); regression practice.
  4. State of GPT | BRK216HFS (42:40) – Challenges.Additional: Pretraining strategies; fine-tuning.

Block 4: Advanced Sequence Modeling

  1. Building makemore Part 5: Building a WaveNet (56:22) – Long-range deps.Additional: CNN basics; graph intros.

Block 5: Transformer Architecture

  1. Let's build GPT: from scratch, in code, spelled out. (1:56:20) – Self-attention.Additional: Multi-head attention; BERT vs. GPT.

Block 6: Tokenization and Data Handling