Key Insights You'll Gain
Understand why LLMs’ capabilities in reasoning, math, and coding emerged later than their language abilities
Learn what Reinforcement Learning with Human Feedback (RLHF) is and why it’s essential for fine-tuning models
Dive into Group Relative Policy Optimization (GRPO), a popular method for optimizing models on logical tasks
Explore practical applications of RLHF and GRPO to make LLMs more reliable in reasoning-intensive scenarios
Meet Our World-Class Speaker
Luis Serrano
Founder, Serrano Academy
Luis Serrano is a renowned AI scientist with a PhD in mathematics and experience in AI at companies that include Google, Apple, and Cohere.