Training Data cover image

Reflection AI’s Misha Laskin on the AlphaGo Moment for LLMs

Training Data

From Narrow to General: The Unlocking Power of Multitasking

1min Snip

00:00
Play full episode
The deep reinforcement learning field initially focused on creating superhuman general agents but ended up producing only narrow agents with high data inefficiency. Training these agents often required excessively large datasets, which posed a challenge for achieving generality. The breakthrough came with the era of language models, which leverage the vast amounts of diverse tasks represented on the internet. By viewing the internet as a collection of various tasks, language models can learn from this multitask environment, leading to enhanced generality in performance.

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode