Machine Learning Street Talk (MLST) cover image

Prof. Randall Balestriero - LLMs without pretraining and SSL

Machine Learning Street Talk (MLST)

00:00

Intro

This chapter examines an experiment with a highly overparameterized model that showcases stable training curves and reduced overfitting. The discussion highlights the implications for language understanding, challenges the necessity of large pre-training datasets, and addresses biases in language models compared to traditional approaches.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app