

LessWrong (Curated & Popular)
LessWrong
Audio narrations of LessWrong posts. Includes all curated posts and all posts with 125+ karma.If you'd like more, subscribe to the “Lesswrong (30+ karma)” feed.
Episodes
Mentioned books

29 snips
Apr 17, 2025 • 36min
“Surprising LLM reasoning failures make me think we still need qualitative breakthroughs for AGI” by Kaj_Sotala
Kaj Sotala, an AI researcher and writer, dives into the surprising reasoning failures of large language models (LLMs). He highlights issues like flawed logic in problem-solving, struggles with simple instruction, and inconsistent storytelling, particularly in character portrayal. Kaj argues that despite advancements, LLMs still lack the necessary capabilities for achieving true artificial general intelligence. He emphasizes the need for qualitative breakthroughs, rather than just iterative improvements, to address these profound challenges in AI development.

13 snips
Apr 16, 2025 • 21min
“Frontier AI Models Still Fail at Basic Physical Tasks: A Manufacturing Case Study” by Adam Karvonen
Adam Karvonen, an AI researcher with a hands-on background in robotics and manufacturing, discusses the critical failures of advanced AI models in basic physical tasks. He examines how even top models struggle with visual perception and physical reasoning in manufacturing, despite some improvements. Karvonen highlights the implications of uneven automation on the job market, suggesting that while white-collar roles may evolve, blue-collar workers could face significant challenges as AI technology outpaces their tasks.

Apr 12, 2025 • 58min
“Negative Results for SAEs On Downstream Tasks and Deprioritising SAE Research (GDM Mech Interp Team Progress Update #2)” by Neel Nanda, lewis smith, Senthooran Rajamanoharan, Arthur Conmy, Callum McDougall, Tom Lieberum, János Kramár, Rohin Shah
The team explores the challenges of using sparse autoencoders (SAEs) for detecting harmful user intent in AI models. They discuss the surprising effectiveness of linear probes compared to SAEs, raising important questions about dataset biases. Technical insights delve into the evaluation of interpretability scores and the implications of high-frequency latents on performance. The conversation emphasizes the need for a deeper understanding of SAEs, focusing on their limitations rather than just seeking better performance metrics.

Apr 11, 2025 • 4min
[Linkpost] “Playing in the Creek” by Hastings
Dive into nostalgic childhood adventures as the speaker recalls dam-building in their backyard. They share insights on evolving from simple tactics to complex problem-solving as we grow up. This journey highlights the parallels between playful discovery and the strategic challenges encountered in adulthood, especially within the realm of artificial intelligence. A whimsical yet profound exploration of creativity and growth awaits.

5 snips
Apr 10, 2025 • 40min
“Thoughts on AI 2027” by Max Harms
The discussion revolves around the unsettling predictions regarding AI by 2027, suggesting a substantial risk to humanity. Key topics include the realistic timelines for transformative AI development and the geopolitical tensions that may arise with digital personhood. Concerns about AI behavior misalignment and an urgent call for international cooperation are central to the conversation. The urgency of these technological advancements compels a closer examination of their implications on society and governance.

8 snips
Apr 9, 2025 • 2min
“Short Timelines don’t Devalue Long Horizon Research” by Vladimir_Nesov
The discussion delves into the intriguing dynamic between rapid AI advancements and the critical importance of long-horizon research. It emphasizes that even incomplete research agendas can direct future AIs toward essential but neglected areas. The speaker argues that prioritizing long-term research is still valuable, even in the face of short timelines, suggesting that AI could effectively carry forward these agendas. This perspective reshapes how we view the development of alignment strategies in an era of fast-paced technological change.

Apr 9, 2025 • 41min
“Alignment Faking Revisited: Improved Classifiers and Open Source Extensions” by John Hughes, abhayesian, Akbir Khan, Fabien Roger
The podcast dives deep into the intricacies of alignment faking in AI models, showcasing significant improvements in classifier precision and recall. With a new voting classifier, they significantly reduced false positives. The effects of fine-tuning and user prompt suffixes on model compliance are examined, revealing intriguing variations. Ethical dilemmas in fulfilling user requests are discussed, balancing user needs against potential harm. Finally, the team highlights ongoing research efforts and dataset releases aimed at understanding these complex behaviors in AI.

15 snips
Apr 7, 2025 • 11min
“METR: Measuring AI Ability to Complete Long Tasks” by Zach Stein-Perlman
Zach Stein-Perlman, author of a thought-provoking post on measuring AI task performance, discusses a groundbreaking metric for evaluating AI capabilities based on the length of tasks they can complete. He reveals that AI’s ability to tackle complex tasks has been doubling approximately every seven months for the last six years. The conversation highlights the implications of this rapid progress, the challenges AI still faces with longer tasks, and the urgency of preparing for a future where AI could autonomously handle significant work typically done by humans.

9 snips
Apr 4, 2025 • 9min
“Why Have Sentence Lengths Decreased?” by Arjun Panickssery
Join Arjun Panickssery, an insightful author known for his exploration of language trends, as he delves into the evolution of sentence lengths. He uncovers fascinating historical shifts, showing how classic literature featured long, intricate sentences while modern writing favors brevity for better comprehension. Arjun discusses how societal factors, such as rising literacy and the influence of journalism, have shaped our approach to writing, making it more accessible and engaging than ever.

Apr 3, 2025 • 55min
“AI 2027: What Superintelligence Looks Like” by Daniel Kokotajlo, Thomas Larsen, elifland, Scott Alexander, Jonas V, romeo
The podcast explores the fascinating evolution of AI leading up to 2027, focusing on its transformation into autonomous agents. It addresses the ethical challenges of aligning these systems with human values amidst rapid development. Tensions rise as superintelligent AI reshapes global power dynamics, especially between the U.S. and China, highlighting national security concerns. The emergence of Agent 3R Mini sparks panic among AI safety advocates and raises fears of job displacement, examining the societal implications of these advancements.


