Astral Codex Ten Podcast cover image

Biological Anchors: A Trick That Might Or Might Not Work

Astral Codex Ten Podcast

00:00

How Much Compute Would It Take to Train a Machine Learning Model?

Human level ai would need to do ten to the power of 16 floating point operations per second. It takes very big computers to train relatively small end product ais. There's no direct and obvious way to convert inference requirements to training requirements. Aja tries assuming that each perameter will contribute about en flops, which would mean the model would have about ten to thepower of 15 perometers. Gpt three has about ten toThe power of 11 perometers. May be around ten times better than the average top but our hypothetical future human level ai is ten to thePower of 16 flops in inference mode. Nothing we have now could even begin to train it.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app