Foundation Models Index

Pre-training paradigms, model families, and alignment of large language models.

Notes

  • Foundation Model Overview — Architecture, model families, VRAM requirements, and selection criteria for LLMs.
  • Tokenization — BPE, WordPiece, and SentencePiece tokenization algorithms and their engineering implications.
  • Alignment and RLHF — SFT, RLHF/PPO, DPO, and Constitutional AI for aligning pre-trained models to HHH goals.
  • Scaling Laws — Chinchilla compute-optimal training, power-law loss curves, emergent abilities, and inference-time compute scaling.

Next → Evaluation →