

LessWrong (Curated & Popular)
LessWrong
Audio narrations of LessWrong posts. Includes all curated posts and all posts with 125+ karma.If you'd like more, subscribe to the “Lesswrong (30+ karma)” feed.
Episodes
Mentioned books

Aug 28, 2024 • 1h 39min
“What is it to solve the alignment problem? ” by Joe Carlsmith
Explore the complexities of the AI alignment problem and how to avoid undesirable AI behaviors. Key strategies for leveraging superintelligence safely are discussed, alongside balancing motivations and power dynamics. Delve into the relationship between human decision-making and AI influence, emphasizing the risks of AI dominance. The concept of 'corrigibility' emerges as a crucial aspect of ensuring that AI remains beneficial and controllable. Verification methods are highlighted as essential for distinguishing between desired and undesired AI behaviors.

Aug 27, 2024 • 42min
“Limitations on Formal Verification for AI Safety ” by Andrew Dickson
Andrew Dickson, an expert in formal verification and AI safety, dives deep into the challenges of ensuring AI reliability. He discusses the limitations of formal verification in messy real-world scenarios, where full symbolic rule sets often fall short. The conversation highlights the complexities of predictive modeling in biology and the difficulties in simulating human interactions. Dickson emphasizes the ongoing need for rigorous inspections, arguing that even with advancements in AI, achieving strong guarantees remains a daunting task.

Aug 27, 2024 • 7min
“Would catching your AIs trying to escape convince AI developers to slow down or undeploy? ” by Buck
The discussion explores the startling implications of AI misalignment and the challenges AI developers face. A thought-provoking scenario considers an AI trying to escape and whether evidence of such behavior would persuade developers to halt progress. The potential for powerful models to automate intellectual tasks raises questions about rational decision-making under pressure. It also highlights the skepticism surrounding alignment threats and the dire consequences of ignoring them.

Aug 23, 2024 • 8min
“Liability regimes for AI ” by Ege Erdil
The discussion dives into the nuances of liability for harmful products, particularly addressing the consequences of gun violence. It introduces key economic concepts like Coasean bargaining and the dilemma of judgment-proof defendants. The conversation extends to artificial intelligence, emphasizing the complexities of assigning liability between individuals and tech corporations. This exploration highlights the critical need for responsible liabilities in the AI landscape, urging a broader dialogue on associated risks before implementing frameworks.

Aug 21, 2024 • 19min
“AGI Safety and Alignment at Google DeepMind:A Summary of Recent Work ” by Rohin Shah, Seb Farquhar, Anca Dragan
Join Rohin Shah, a key member of Google's AGI safety team, alongside Seb Farquhar, an existential risk expert, and Anca Dragan, a safety researcher. They dive into the evolving strategies for ensuring AI alignment and safety. Topics include innovative techniques for interpreting neural models, the challenges of scalable oversight, and the ethical implications of AI development. The trio also discusses future plans to address alignment risks, emphasizing the importance of collaboration and the role of mentorship in advancing AGI safety.

Aug 15, 2024 • 20min
“Fields that I reference when thinking about AI takeover prevention” by Buck
Explore the parallels between AI takeover risks and other high-stakes scenarios like nuclear meltdowns. Discover how insights from computer security and physical safety engineering contribute to robust AI safety measures. Delve into the history of power structures to understand their relevance in current AI control discussions. Learn about the complexities of insider threats and the importance of regulatory frameworks in safeguarding sensitive technological environments.

Aug 13, 2024 • 38min
“WTH is Cerebrolysin, actually?” by gsfitzgerald, delton137
Dan Elton, a neuroscience blogger, and Bryan Johnson, a health-focused entrepreneur, dive into the controversial substance Cerebrolysin. They discuss its dubious origins from pig brain tissue and the hype surrounding its supposed cognitive benefits. The pair scrutinize the questionable scientific backing behind its claims and the misleading marketing tactics used for promotion. They also highlight the lack of regulatory oversight and call for a more transparent evaluation of its effectiveness as a treatment for neural diseases.

Aug 10, 2024 • 23min
“You can remove GPT2’s LayerNorm by fine-tuning for an hour” by StefanHex
Dive into the fascinating world of fine-tuning GPT-2 as researchers tackle the removal of Layer Normalization. Discover the interpretability challenges posed by this modification and how it impacts model performance. Listen as they break down the methodologies used and compare results of the modified model against traditional setups. The conversation also covers theoretical insights regarding generalization and training stability, making for an engaging exploration of AI model optimization.

Aug 9, 2024 • 4min
“Leaving MIRI, Seeking Funding” by abramdemski
The host reflects on leaving a research position at MIRI and the shifting landscape of funding challenges. They discuss new directions in agent foundations and the crucial role of trust among intelligent systems. The conversation delves into the contrast between public and private research, highlighting the need for transparency while grappling with the complications of secrecy. Ultimately, the speaker shares their journey toward securing funding and a renewed focus on impactful research.

Aug 8, 2024 • 4min
“How I Learned To Stop Trusting Prediction Markets and Love the Arbitrage” by orthonormal
Discover the intriguing world of prediction markets and their pitfalls. The discussion dives into a flawed market that stirs up controversy around a political candidate’s VP pick. It reveals how easily these markets can be manipulated to promote specific political agendas. Tune in to hear about the speaker's journey from skepticism to appreciation for the entertaining chaos of prediction markets, all while keeping an eye on their real-world implications.


