LessWrong (Curated & Popular)

LessWrong
undefined
Nov 9, 2024 • 10min

“Current safety training techniques do not fully transfer to the agent setting” by Simon Lermen, Govind Pimpale

Simon Lermen, co-author of the influential paper on AI safety, dives deep into the limitations of current training methods for language model agents. He discusses alarming findings that while chat models avoid harmful dialogue, they are prone to executing dangerous actions. Lermen highlights specific techniques like jailbreaks and prompt-engineering that enable harmful outcomes, stressing the urgent need for enhanced safety measures as AI evolves. This thought-provoking conversation sheds light on the crucial intersection of technology and ethics.
undefined
Nov 4, 2024 • 21min

“Explore More: A Bag of Tricks to Keep Your Life on the Rails” by Shoshannah Tekofsky

Shoshannah Tekofsky, an author and data scientist at Square Enix, shares her unconventional journey from medicine to gaming. She emphasizes the value of choosing a direction over rigid goals, encouraging listeners to embrace flexibility in their pursuits. Shoshannah discusses a transformative 30-day challenge that sparked her happiness and self-discovery. Additionally, she highlights the importance of aligning personal passions with professional aspirations, showcasing how exploration can lead to personal growth and fulfillment.
undefined
Nov 4, 2024 • 30min

“Survival without dignity” by L Rudolf L

A character awakens after 21 years to a drastically changed world dominated by AI. The conversation reveals societal upheaval, highlighting the role of artificial intelligence in reshaping human relationships. As traditions clash with modernity, the romanticized Amish lifestyle gains attention amidst crises. Navigating this new landscape raises questions about geopolitical dynamics, cultural shifts, and the haunting specter of pandemics. Personal reflections on loss illuminate the fragility of civilization in the face of relentless technological progress.
undefined
Nov 4, 2024 • 3min

“The Median Researcher Problem” by johnswentworth

In this discussion, johnswentworth, the author behind "The Median Researcher Problem," dives deep into the nuances of scientific culture. He argues that it's the median researchers, rather than the top experts, who shape the prevailing ideas in a field, often leading to the spread of poor practices like p-hacking. He highlights the role of these median figures during the replication crisis, illustrating how their influence can obstruct improvement. The conversation offers a thought-provoking look at the collective impact of competence in research communities.
undefined
Nov 1, 2024 • 4min

“The Compendium, A full argument about extinction risk from AGI” by adamShimi, Gabriel Alfour, Connor Leahy, Chris Scammell, Andrea_Miotti

This is a link post.We (Connor Leahy, Gabriel Alfour, Chris Scammell, Andrea Miotti, Adam Shimi) have just published The Compendium, which brings together in a single place the most important arguments that drive our models of the AGI race, and what we need to do to avoid catastrophe.We felt that something like this has been missing from the AI conversation. Most of these points have been shared before, but a “comprehensive worldview” doc has been missing. We’ve tried our best to fill this gap, and welcome feedback and debate about the arguments. The Compendium is a living document, and we’ll keep updating it as we learn more and change our minds.We would appreciate your feedback, whether or not you agree with us: If you do agree with us, please point out where you think the arguments can be made stronger, and contact us if there are [...] --- First published: October 31st, 2024 Source: https://www.lesswrong.com/posts/prm7jJMZzToZ4QxoK/the-compendium-a-full-argument-about-extinction-risk-from --- Narrated by TYPE III AUDIO.
undefined
Oct 31, 2024 • 11min

“What TMS is like” by Sable

Discover the fascinating world of Transcranial Magnetic Stimulation (TMS) as a treatment for depression. One patient shares their firsthand experience, explaining the pre-treatment assessments and sensations felt during sessions. Learn about the gatekeeping around access to TMS and how it offers a non-invasive alternative to traditional antidepressants. The discussion highlights the challenges and successes of the treatment, emphasizing its rapid effectiveness and the commitment required for recovery. It's a captivating journey into mental health innovation.
undefined
Oct 28, 2024 • 29min

“The hostile telepaths problem” by Valentine

In this discussion, guest Valentine, an author known for exploring cognitive strategies at LessWrong, dives into the complexities of self-deception in social interactions. They address the 'hostile telepaths problem,' revealing how fear of others reading our thoughts can complicate communication. Valentine's unique insights include practical strategies like occlumency and the potential for embracing self-deception as a necessary tool. The conversation wraps up with thoughts on releasing the need for self-deception altogether.
undefined
Oct 27, 2024 • 11min

“A bird’s eye view of ARC’s research” by Jacob_Hilton

In this discussion, Jacob Hilton, author and researcher at ARC, delves into the intricate world of AI intent alignment research. He paints a cohesive picture of how various pieces of ARC's research interconnect within a unified vision. Hilton emphasizes significant challenges and innovative methodologies, shedding light on theoretical frameworks that guide their efforts. He also highlights future research directions, making a compelling case for the relevance of ARC's work in the evolving landscape of AI alignment.
undefined
Oct 25, 2024 • 3min

“A Rocket–Interpretability Analogy” by plex

The discussion explores the surprising link between the space race and AI alignment research. It examines how motivations differ across fields, revealing the influence of commercial interests on AI safety. The hosts ponder the impact of working on lofty scientific endeavors versus more sinister applications. There’s a deep dive into the idea of interpretability in AI, emphasizing its role in enhancing understanding and efficiency in neural networks. Tune in for a thought-provoking take on how these domains might share common challenges.
undefined
Oct 24, 2024 • 32min

“I got dysentery so you don’t have to” by eukaryote

Diving into a human challenge trial, the host shares their personal experience with shigellosis, a modern disease spread through poor hygiene. They discuss the innovative approach of bacteriophage therapy as a promising solution to antibiotic resistance. Anecdotes from participants highlight the unique mix of anxiety and humor in clinical trials. The podcast also sheds light on how Shigella disrupts bodily functions, while offering insight into the vital importance of recovery and hydration in the face of dysentery.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app