80,000 Hours Podcast cover image

#107 – Chris Olah on what the hell is going on inside neural networks

80,000 Hours Podcast

00:00

An Introduction to Scaling Laws in Model Performance

Scaling laws, represented by straight lines on log-log plots, are of great interest. The most important scaling law is the relationship between model size and performance loss. As model size increases, loss decreases. This follows a power law equation, where one variable is raised to a power. People are finding that increasing model size does not result in diminishing returns as expected. Performance continues to improve at a steady pace. However, it is uncertain if this trend will continue indefinitely. The straight line pattern on the plot is both surprising and tempting to extrapolate. It is possible to reason about models larger than those currently trainable.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app