
Exhaustion of High-Quality Data Could Slow Down AI Progress in Coming Decades
The Data Exchange with Ben Lorica
00:00
Scaling Loss and the Importance of Data
The more data you have, the better your model is. But for transfer learning to other tasks, then the relationship becomes a bit more blurry. At some points, you need diminishing returns and you approach an asymptotic performance when you have infinite data if you need compute. This kind of is goodness for the previous paper, because at some point, then you don't need additional data. Well, I would actually say it's bad news, because if you want the same improvement that you had from in a given year, then you need to increase the amount of data even more next year.
Transcript
Play full episode