Building Attention from Scratch
Implement the attention mechanism that powers transformers, from matrix operations to multi-head attention, with clear PyTorch code.
Deep technical explorations of AI concepts, complete with production-ready code
Implement the attention mechanism that powers transformers, from matrix operations to multi-head attention, with clear PyTorch code.
Build a miniature GPT model that actually generates coherent text. Covers tokenization, positional encoding, and training loops.
Why do LLMs represent words as vectors? Learn the intuition behind embeddings and build your own word2vec implementation.
Derive backpropagation from calculus first principles. Build a neural network training engine with automatic differentiation.
Deploy language models in production: quantization, batching, caching strategies, and monitoring for real-world applications.
Build CNNs without frameworks. Understand convolutions, pooling, and feature maps by implementing them with NumPy.
How do LLMs break text into pieces? Build BPE tokenization from scratch and understand subword algorithms.
Adapt foundation models to your domain. LoRA, prompt tuning, and full fine-tuning with practical examples.
Visualize how neural networks learn. Implement SGD, Adam, and learning rate schedules from mathematical foundations.