LessWrong (Curated & Popular)

LessWrong
undefined
Aug 28, 2024 • 1h 39min

“What is it to solve the alignment problem? ” by Joe Carlsmith

Explore the complexities of the AI alignment problem and how to avoid undesirable AI behaviors. Key strategies for leveraging superintelligence safely are discussed, alongside balancing motivations and power dynamics. Delve into the relationship between human decision-making and AI influence, emphasizing the risks of AI dominance. The concept of 'corrigibility' emerges as a crucial aspect of ensuring that AI remains beneficial and controllable. Verification methods are highlighted as essential for distinguishing between desired and undesired AI behaviors.
undefined
Aug 27, 2024 • 42min

“Limitations on Formal Verification for AI Safety ” by Andrew Dickson

Andrew Dickson, an expert in formal verification and AI safety, dives deep into the challenges of ensuring AI reliability. He discusses the limitations of formal verification in messy real-world scenarios, where full symbolic rule sets often fall short. The conversation highlights the complexities of predictive modeling in biology and the difficulties in simulating human interactions. Dickson emphasizes the ongoing need for rigorous inspections, arguing that even with advancements in AI, achieving strong guarantees remains a daunting task.
undefined
Aug 27, 2024 • 7min

“Would catching your AIs trying to escape convince AI developers to slow down or undeploy? ” by Buck

The discussion explores the startling implications of AI misalignment and the challenges AI developers face. A thought-provoking scenario considers an AI trying to escape and whether evidence of such behavior would persuade developers to halt progress. The potential for powerful models to automate intellectual tasks raises questions about rational decision-making under pressure. It also highlights the skepticism surrounding alignment threats and the dire consequences of ignoring them.
undefined
Aug 23, 2024 • 8min

“Liability regimes for AI ” by Ege Erdil

The discussion dives into the nuances of liability for harmful products, particularly addressing the consequences of gun violence. It introduces key economic concepts like Coasean bargaining and the dilemma of judgment-proof defendants. The conversation extends to artificial intelligence, emphasizing the complexities of assigning liability between individuals and tech corporations. This exploration highlights the critical need for responsible liabilities in the AI landscape, urging a broader dialogue on associated risks before implementing frameworks.
undefined
Aug 21, 2024 • 19min

“AGI Safety and Alignment at Google DeepMind:A Summary of Recent Work ” by Rohin Shah, Seb Farquhar, Anca Dragan

Join Rohin Shah, a key member of Google's AGI safety team, alongside Seb Farquhar, an existential risk expert, and Anca Dragan, a safety researcher. They dive into the evolving strategies for ensuring AI alignment and safety. Topics include innovative techniques for interpreting neural models, the challenges of scalable oversight, and the ethical implications of AI development. The trio also discusses future plans to address alignment risks, emphasizing the importance of collaboration and the role of mentorship in advancing AGI safety.
undefined
Aug 15, 2024 • 20min

“Fields that I reference when thinking about AI takeover prevention” by Buck

Explore the parallels between AI takeover risks and other high-stakes scenarios like nuclear meltdowns. Discover how insights from computer security and physical safety engineering contribute to robust AI safety measures. Delve into the history of power structures to understand their relevance in current AI control discussions. Learn about the complexities of insider threats and the importance of regulatory frameworks in safeguarding sensitive technological environments.
undefined
Aug 13, 2024 • 38min

“WTH is Cerebrolysin, actually?” by gsfitzgerald, delton137

Dan Elton, a neuroscience blogger, and Bryan Johnson, a health-focused entrepreneur, dive into the controversial substance Cerebrolysin. They discuss its dubious origins from pig brain tissue and the hype surrounding its supposed cognitive benefits. The pair scrutinize the questionable scientific backing behind its claims and the misleading marketing tactics used for promotion. They also highlight the lack of regulatory oversight and call for a more transparent evaluation of its effectiveness as a treatment for neural diseases.
undefined
Aug 10, 2024 • 23min

“You can remove GPT2’s LayerNorm by fine-tuning for an hour” by StefanHex

Dive into the fascinating world of fine-tuning GPT-2 as researchers tackle the removal of Layer Normalization. Discover the interpretability challenges posed by this modification and how it impacts model performance. Listen as they break down the methodologies used and compare results of the modified model against traditional setups. The conversation also covers theoretical insights regarding generalization and training stability, making for an engaging exploration of AI model optimization.
undefined
Aug 9, 2024 • 4min

“Leaving MIRI, Seeking Funding” by abramdemski

The host reflects on leaving a research position at MIRI and the shifting landscape of funding challenges. They discuss new directions in agent foundations and the crucial role of trust among intelligent systems. The conversation delves into the contrast between public and private research, highlighting the need for transparency while grappling with the complications of secrecy. Ultimately, the speaker shares their journey toward securing funding and a renewed focus on impactful research.
undefined
Aug 8, 2024 • 4min

“How I Learned To Stop Trusting Prediction Markets and Love the Arbitrage” by orthonormal

Discover the intriguing world of prediction markets and their pitfalls. The discussion dives into a flawed market that stirs up controversy around a political candidate’s VP pick. It reveals how easily these markets can be manipulated to promote specific political agendas. Tune in to hear about the speaker's journey from skepticism to appreciation for the entertaining chaos of prediction markets, all while keeping an eye on their real-world implications.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app