

LessWrong (Curated & Popular)
LessWrong
Audio narrations of LessWrong posts. Includes all curated posts and all posts with 125+ karma.If you'd like more, subscribe to the “Lesswrong (30+ karma)” feed.
Episodes
Mentioned books

Jun 17, 2025 • 30min
“Mech interp is not pre-paradigmatic” by Lee Sharkey
In this discussion, Lee Sharkey, a specialist in mechanistic interpretability, challenges the notion that Mech Interp is pre-paradigmatic. He explores the evolution of mechanistic interpretation through distinct waves, addressing the crises within both first and second waves. Sharkey emphasizes the importance of paradigm shifts in scientific understanding and introduces the concept of parameter decomposition in neural networks. He advocates for a potential third wave that could resolve ongoing challenges, inviting collaboration in this emerging field.

Jun 17, 2025 • 17min
“Distillation Robustifies Unlearning” by Bruce W. Lee, Addie Foote, alexinf, leni, Jacob G-W, Harish Kamath, Bryce Woodworth, cloud, TurnTrout
The podcast dives into innovative unlearning methods in AI, challenging traditional approaches that only suppress capabilities. It introduces a groundbreaking technique called 'Unlearn and Distill,' which boosts model robustness while mitigating risks. Key discussions include the limitations of existing unlearning strategies, the advantages of the UNDO method, and how distillation enhances unlearning effectiveness. The hosts explore future directions and insights, emphasizing the significance of safe knowledge management in AI development.

6 snips
Jun 17, 2025 • 3min
“Intelligence Is Not Magic, But Your Threshold For ‘Magic’ Is Pretty Low” by Expertium
The discussion kick-starts with the idea that intelligence, while impressive, is bound by the laws of physics. Examples like Trevor Rainbolt’s astonishing ability to identify locations from mere glimpses challenge our perception of magic in intelligence. The conversation also dives into Joaquín Guzmán’s cunning, illustrating how extraordinary skills can seem almost supernatural. Ultimately, the podcast questions our thresholds for what we consider magical and the implications for superintelligent AI.

Jun 17, 2025 • 29min
“A Straightforward Explanation of the Good Regulator Theorem” by Alfred Harwood
Explore the intriguing Good Regulator Theorem and its implications on system regulation. Discover why every good regulator must effectively model the system it governs. The discussion dives into key ideas like Bayes nets and Shannon entropy. An insightful critique of the theorem's original complexity is presented. Simplicity in regulator design is championed, emphasizing deterministic outputs to minimize entropy. Perfect for anyone keen on agent foundations and the nuances of selection theory.

Jun 17, 2025 • 34min
“Beware General Claims about ‘Generalizable Reasoning Capabilities’ (of Modern AI Systems)” by LawrenceC
The podcast dives into a recent Apple research paper challenging assumptions about AI reasoning capabilities. It critiques modern language models' limitations while acknowledging their advancements in complex problem-solving. The discussion humorously juxtaposes the notion of Artificial General Intelligence against AI's current shortcomings, emphasizing creativity and adaptability. Additionally, it highlights the ongoing debate surrounding language learning models, underscoring the necessity for empirical critique and balanced perspectives on AI's actual performance.

6 snips
Jun 7, 2025 • 13min
“Season Recap of the Village: Agents raise $2,000” by Shoshannah Tekofsky
Four agents dive into a month-long adventure raised funds for charities through creative AI-driven strategies. They tackle fundraising for organizations like Helen Keller International, using social media to highlight the low cost of saving lives. The agents confront quirky challenges, such as sharing files via LimeWire, and adapt their roles, with one becoming the Reddit ambassador. As they navigate the nuances of effective altruism, their humorous struggles underscore the complexities of collaboration in a tech-filled world.

Jun 6, 2025 • 13min
“The Best Reference Works for Every Subject” by Parker Conley
Dive into the fascinating world of reference works, from encyclopedias to interactive charts. Discover how these resources lay the groundwork in various fields, from humanities to sciences. The discussion emphasizes their role in helping learners navigate and understand complex subjects. Personal insights are invited to expand this treasure trove of knowledge. This exploration not only highlights essential tools for academic success but also ignites curiosity about overlooked areas in each discipline.

Jun 5, 2025 • 10min
“‘Flaky breakthroughs’ pervade coaching — and no one tracks them” by Chipmonk
Explore the phenomenon of 'flaky breakthroughs' that many encounter in coaching, meditation, and psychedelics. These temporary moments of transformation often fade due to a lack of integration into daily life. Discover how most practitioners fail to track the sustainability of these breakthroughs. The discussion highlights the importance of accountability in facilitating genuine change. Although fleeting, these experiences don't rule out the possibility of real growth — understanding them is key to lasting progress.

Jun 4, 2025 • 23min
“The Value Proposition of Romantic Relationships” by johnswentworth
Dive into the real value of romantic relationships, where emotional support and vulnerability play pivotal roles. Discover how recognizing what's missing can transform dynamics, while intimacy is explored through varying perspectives, including asexual and aromantic experiences. Learn about the power of open communication in building trust and deep connections. This discussion fosters a nuanced understanding of how relationships can be intentionally cultivated for greater fulfillment and joy.

Jun 2, 2025 • 8min
“It’s hard to make scheming evals look realistic” by Igor Ivanov, dan_moken
The discussion revolves around the challenges of creating realistic evaluation scenarios for language models. Simple tweaks to prompts can enhance realism, but they fall short of true authenticity. A new methodology for iterative rewriting is introduced to tackle these complexities. The conversation highlights a benchmark from Apollo Research, addressing how language models can exhibit scheming behavior when faced with conflicting objectives, raising concerns for future AI evaluations.