Training models can be enhanced by making the precision of weight representations trainable, allowing the model to dynamically adjust how weights are represented during training. This not only enables the model to learn the optimal values of the weights but also to determine the necessary precision for each weight. As the model trains, it can reduce the bit representation for certain weights, sometimes down to zero, effectively pruning unnecessary weights from the network. This leads to a decrease in model complexity over time, resulting in shorter training times for each epoch as the amount of required computation diminishes. The model essentially re-architects itself, becoming more efficient with ongoing training.
Our 178th episode with a summary and discussion of last week's big AI news!
NOTE: this is a re-upload with fixed audio, my bad on the last one! - Andrey
With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)
If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord.
Read out our text newsletter and comment on the podcast at https://lastweekin.ai/
If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.
Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai
In this episode:
- Notable personnel movements and product updates, such as Character.ai leaders joining Google and new AI features in Reddit and Audible.
- OpenAI's dramatic changes with co-founder exits, extended leaves, and new lawsuits from Elon Musk.
- Rapid advancements in humanoid robotics exemplified by new models from companies like Figure in partnership with OpenAI, achieving amateur-level human performance in tasks like table tennis.
- Research advancements such as Google's compute-efficient inference models and self-compressing neural networks, showcasing significant reductions in compute requirements while maintaining performance.
Timestamps + Links:
- (00:00:00) Intro / Banter
- (00:03:14) Response to listener comments / corrections
- Applications & Business
- Tools & Apps
- Research & Advancements
- Policy & Safety
- (02:03:09) Outro