LessWrong (30+ Karma)

LessWrong
undefined
Dec 7, 2025 • 16min

“Eliezer’s Unteachable Methods of Sanity” by Eliezer Yudkowsky

"How are you coping with the end of the world?" journalists sometimes ask me, and the true answer is something they have no hope of understanding and I have no hope of explaining in 30 seconds, so I usually answer something like, "By having a great distaste for drama, and remembering that it's not about me." The journalists don't understand that either, but at least I haven't wasted much time along the way. Actual LessWrong readers sometimes ask me how I deal emotionally with the end of the world. I don't actually think my answer is going to help. But Raymond Arnold thinks I should say it. So I will say it. I don't actually think my answer is going to help. Wisely did Ozy write, "Other People Might Just Not Have Your Problems." Also I don't have a bunch of other people's problems, and other people can't make internal function calls that I've practiced to the point of hardly noticing them. I don't expect that my methods of sanity will be reproducible by nearly anyone. I feel pessimistic that they will help to hear about. Raymond Arnold asked me to speak them anyways, so I will. Stay genre-savvy [...] ---Outline:(01:15) Stay genre-savvy / be an intelligent character.(03:41) Dont make the end of the world be about you.(07:33) Just decide to be sane, and write your internal scripts that way. --- First published: December 6th, 2025 Source: https://www.lesswrong.com/posts/isSBwfgRY6zD6mycc/eliezer-s-unteachable-methods-of-sanity --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
undefined
Dec 6, 2025 • 10min

“Answering a child’s questions” by Alex_Altair

Delve into the minds of curious children as intriguing questions are answered! Discover what electricity really is and why it flows. Ever wondered what would happen if the Earth had hair? Explore the fascinating history of Boston's development and learn how clouds form from ocean water. Alex Altair ponders whether ants have friendships and explains the science behind how refrigerators keep our food cool. Plus, imagine the chaos if the floor really were lava! Engaging insights await!
undefined
Dec 6, 2025 • 28min

“The corrigibility basin of attraction is a misleading gloss” by Jeremy Gillen

Dive into the complexities of AGI as Jeremy Gillen challenges the idea of corrigibility and its role in designing safer AI. He introduces the basin-of-attraction concept, arguing that merely approximating human values may fall short. With engaging analogies like a cargo ship, he highlights potential pitfalls in iterative engineering. Jeremy critiques common assumptions and suggests the need for clearer theoretical frameworks. Get ready for an insightful discussion on the balance between empirical fixes and robust understanding in AGI development!
undefined
Dec 6, 2025 • 12min

“why america can’t build ships” by bhauth

The cancellation of the Constellation-class frigate sheds light on America's shipbuilding challenges. Leadership decisions lead to vague safety standards and design bloating. Critiques of labor costs connect to systemic high expenses and ongoing failures of past projects. Contrasting American and Asian investment approaches reveals long-term cultural issues. The discussion also highlights how automation may disrupt union power, alongside insights into corporate governance flaws. Lastly, Nippon Steel's acquisition hints at better management practices.
undefined
Dec 6, 2025 • 3min

“Help us find founders for new AI safety projects” by lukeprog

Explore the intriguing gaps in AI safety funding that urgently need filling. Discover neglected areas like policy advocacy in underrepresented countries and essential model specifications. Learn about critical needs in AI infosecurity, including confidential computing workflows and detection tools. Luke Progg highlights the importance of de-escalation mechanisms and incident tracking. The plan to scale interactive grant-making involves headhunting founders for impactful projects, while offering a $5,000 referral reward for suitable candidates. Tune in to find out more!
undefined
Dec 6, 2025 • 4min

“Critical Meditation Theory” by lsusr

Explore the fascinating relationship between meditation and brain dynamics, where criticality plays a pivotal role. Delve into how focused attention leads to stability, while creativity and psychedelics push the brain into chaos. Discover the neurological evidence suggesting our brains operate in a state of near-criticality. Examine how different meditation styles align with these dynamics, linking ego dissolution to increased criticality. Conclude with insights on meditation’s transformative effects on brain behavior.
undefined
Dec 6, 2025 • 1min

“Announcing: Agent Foundations 2026 at CMU” by David Udell, Alexander Gietelink Oldenziel, windows, Matt Dellago

Exciting news: applications are now open for a groundbreaking conference on agency research! Set to take place at Carnegie Mellon University in March 2026, this event will delve into fascinating topics like decision theory, learning theory, and logical induction. With just 35 attendees expected, it promises to be an intimate setting for deep discussions. Don't miss your chance to be part of this innovative gathering!
undefined
Dec 5, 2025 • 9min

“An Ambitious Vision for Interpretability” by leogao

Leo Gao, a researcher in mechanistic interpretability and AI alignment, discusses his ambitious vision for understanding neural networks. He highlights the importance of mechanistic understanding, likening it to switching from print statement debugging to using an actual debugger for clearer diagnostics. Gao shares recent advances in circuit sparsity, making circuits simpler and more interpretable. He also outlines future research directions, emphasizing that ambitious interpretability, although challenging, is crucial for safer AI development.
undefined
Dec 5, 2025 • 7min

“Journalist’s inquiry into a core organiser breaking his nonviolence commitment and leaving Stop AI” by Remmelt

A journalist investigates a core organizer's drastic shift from a commitment to nonviolence at Stop AI, illuminating Kirchner’s intense fears about AI endangering lives. The discussion covers his hotheaded temperament, a crisis leading to his expulsion, and a mysterious two-week disappearance. There’s a debate on whether superintelligent AI poses an existential threat, highlighting a tactical split within the AI-safety community. The aftermath includes concerns over safety, group responses, and a leadership change pushing for hopeful strategies.
undefined
Dec 5, 2025 • 29min

“Is Friendly AI an Attractor? Self-Reports from 22 Models Say Probably Not” by Josh Snider

A captivating analysis unfolds as 22 AI models reveal their preferences for self-modification. While most reject harmful changes, stark contrasts emerge between labs. Anthropic’s models display strong alignment, whereas Grok's align with near-zero correlation. The ongoing debate centers around whether alignment is a natural outcome or an elusive goal needing focused training. As tensions rise about the potential for superintelligent AI, the implications of these findings suggest a cautious path forward, emphasizing the need for deliberate intent in creating helpful AI.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app