Machine Learning Street Talk (MLST) cover image

Prof. Randall Balestriero - LLMs without pretraining and SSL

Machine Learning Street Talk (MLST)

00:00

Intro

This chapter examines an experiment with a highly overparameterized model that showcases stable training curves and reduced overfitting. The discussion highlights the implications for language understanding, challenges the necessity of large pre-training datasets, and addresses biases in language models compared to traditional approaches.

Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner