How LLMs Work: Reinforcement Learning, RLHF, DeepSeek R1, OpenAI o1, AlphaGo
Part 2 of the LLM deep dive The post How LLMs Work: Reinforcement Learning, RLHF, DeepSeek R1, OpenAI o1, AlphaGo appeared first on Towards Data Science.

Welcome to part 2 of my LLM deep dive. If you’ve not read Part 1, I highly encourage you to check it out first.
Previously, we covered the first two major stages of training an LLM:
- Pre-training — Learning from massive datasets to form a base model.
- Supervised fine-tuning (SFT) — Refining the model with curated examples to make it useful.
Now, we’re diving into the next major stage: Reinforcement Learning (RL). While pre-training and SFT are well-established, RL is still evolving but has become a critical part of the training pipeline.
I’ve taken reference from Andrej Karpathy’s widely popular 3.5-hour YouTube. Andrej is a founding member of OpenAI, his insights are gold — you get the idea.
Let’s go
Read More