The MAD Podcast with Matt Turck cover image

Open Source AI Strikes Back — Inside Ai2’s OLMo 3 ‘Thinking"

The MAD Podcast with Matt Turck

00:00

RLVR: Reinforcement with Verifiable Rewards

Nathan outlines RLVR vs RLHF: using verifiable correctness (e.g., tests) as rewards for stable optimization.

Play episode from 01:10:51
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app