Reasoning and Alignment in Large Language Models
This module explores how language models transcend simple pattern prediction to develop genuine reasoning capabilities and align with human values through reinforcement learning.
Why language models need reinforcement learning and how it enables new capabilities
How reinforcement learning from human feedback transforms text predictors into helpful assistants
How models develop sophisticated reasoning through chain-of-thought and test-time computation
Synthesis of the entire evolution from n-grams to sophisticated reasoning systems