Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work!
Visit our Patreon page to unlock exclusive Bayesian swag ;)
Takeaways:
- The hype around AI in science often fails to deliver practical results.
- Bayesian deep learning combines the strengths of deep learning and Bayesian statistics.
- Fine-tuning LLMs with Bayesian methods improves prediction calibration.
- There is no single dominant library for Bayesian deep learning yet.
- Real-world applications of Bayesian deep learning exist in various fields.
- Prior knowledge is crucial for the effectiveness of Bayesian deep learning.
- Data efficiency in AI can be enhanced by incorporating prior knowledge.
- Generative AI and Bayesian deep learning can inform each other.
- The complexity of a problem influences the choice between Bayesian and traditional deep learning.
- Meta-learning enhances the efficiency of Bayesian models.
- PAC-Bayesian theory merges Bayesian and frequentist ideas.
- Laplace inference offers a cost-effective approximation.
- Subspace inference can optimize parameter efficiency.
- Bayesian deep learning is crucial for reliable predictions.
- Effective communication of uncertainty is essential.
- Realistic benchmarks are needed for Bayesian methods
- Collaboration and communication in the AI community are vital.
Chapters:
00:00 Introduction to Bayesian Deep Learning
04:24 Vincent Fortuin’s Journey to Bayesian Deep Learning
11:52 Understanding Bayesian Deep Learning
16:29 Current Landscape of Bayesian Libraries
21:11 Real-World Applications of Bayesian Deep Learning
23:33 When to Use Bayesian Deep Learning
28:22 Data Efficiency in AI and Generative Modeling
30:18 Integrating Bayesian Knowledge into Generative Models
31:44 The Role of Meta-Learning in Bayesian Deep Learning
34:06 Understanding Pack Bayesian Theory
37:55 Algorithms for Bayesian Deep Learning Models
42:10 Advancements in Efficient Inference Techniques
48:21 The Importance of Uncertainty in AI Models
51:41 Advice for Aspiring Researchers in AI
54:08 Future Directions in AI for Science
Thank you to my Patrons for making this episode possible!
Yusuke Saito, Avi Bryant, Ero Carrera, Giuliano Cruz, Tim Gasser, James Wade, Tradd Salvo, William Benton, James Ahloy, Robin Taylor,, Chad Scherrer, Zwelithini Tunyiswa, Bertrand Wilden, James Thompson, Stephen Oates, Gian Luca Di Tanna, Jack Wells, Matthew Maldonado, Ian Costley, Ally Salim, Larry Gill, Ian Moran, Paul Oreto, Colin Caprani, Colin Carroll, Nathaniel Burbank, Michael Osthege, Rémi Louf, Clive Edelsten, Henri Wallen, Hugo Botha, Vinh Nguyen, Marcin Elantkowski, Adam C. Smith, Will Kurt, Andrew Moskowitz, Hector Munoz, Marco Gorelli, Simon Kessell, Bradley Rode, Patrick Kelley, Rick Anderson, Casper de Bruin, Philippe Labonde, Michael Hankin, Cameron Smith, Tomáš Frýda, Ryan Wesslen, Andreas Netti, Riley King, Yoshiyuki Hamajima, Sven De Maeyer, Michael DeCrescenzo, Fergal M, Mason Yahr, Naoya Kanai, Steven Rowland, Aubrey Clayton, Jeannine Sue, Omri Har Shemesh, Scott Anthony Robson, Robert Yolken, Or Duek, Pavel Dusek, Paul Cox, Andreas Kröpelin, Raphaël R, Nicolas Rode, Gabriel Stechschulte, Arkady, Kurt TeKolste, Gergely Juhasz, Marcus Nölke, Maggi Mackintosh, Grant Pezzolesi, Avram Aelony, Joshua Meehl, Javier Sabio, Kristian Higgins, Alex Jones, Gregorio Aguilar, Matt Rosinski, Bart Trudeau, Luis Fonseca, Dante Gates, Matt Niccolls, Maksim Kuznecov, Michael Thomas, Luke Gorrie, Cory Kiser, Julio, Edvin Saveljev, Frederick Ayala, Jeffrey Powell, Gal Kampel, Adan Romero, Will Geary, Blake Walters, Jonathan Morgan, Francesco Madrisotti, Ivy Huang, Gary Clarke, Robert Flannery, Rasmus Hindström, Stefan, Corey Abshire, Mike Loncaric, David McCormick, Ronald Legere, Sergio Dolia, Michael Cao, Yiğit Aşık and Suyog Chandramouli.
Links from the show:
Transcript
This is an automatic transcript and may therefore contain errors. Please get in touch if you're willing to correct them.