3min chapter

The Data Exchange with Ben Lorica cover image

Exhaustion of High-Quality Data Could Slow Down AI Progress in Coming Decades

The Data Exchange with Ben Lorica

CHAPTER

Scaling Loss and the Importance of Data

The more data you have, the better your model is. But for transfer learning to other tasks, then the relationship becomes a bit more blurry. At some points, you need diminishing returns and you approach an asymptotic performance when you have infinite data if you need compute. This kind of is goodness for the previous paper, because at some point, then you don't need additional data. Well, I would actually say it's bad news, because if you want the same improvement that you had from in a given year, then you need to increase the amount of data even more next year.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode