2min chapter

LessWrong (Curated & Popular) cover image

"Two-year update on my personal AI timelines" by Ajeya Cotra

LessWrong (Curated & Popular)

CHAPTER

Explicitly Breaking Out GPT-N as an Anchor

Short horizon inefficiently trained coding models operating pretty close to their training distributions of massively accelerated AI research. I'm now explicitly putting significant weight on an amount of compute that's more like just scaling up language models to brainish sizes. This is consistent with doing RL fine tuning, but just needing many OOMs less data for that than for the original training run. And I think that's the most likely way it would manifest.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode