2min chapter

The Gradient: Perspectives on AI cover image

Zachary Lipton: Where Machine Learning Falls Short

The Gradient: Perspectives on AI

CHAPTER

Pre Training for Summarization

There's a question of, is it that this is a useful routine that transformers are just somehow unstable or ill conditioned? Right. And actually, you know, we have a new paper that I'm kind of excited about that I can tell you that should be on the archive tonight. So for summarization, people tend to use these like things where this pre training objects for everything is represented as a sequence like T5 types that up. For classification people might use a BERT model, signing train or the mass language modeling objective and they're doing this with larger and larger transformer models,. Not larger and larger core poor and seeing over getting these gains. But there's a question which is

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode