LessWrong (Curated & Popular) cover image

LessWrong (Curated & Popular)

Latest episodes

undefined
May 5, 2025 • 13min

“Interpretability Will Not Reliably Find Deceptive AI” by Neel Nanda

Neel Nanda, a thought leader on AI safety, shares his intriguing insights on interpretability and its limits. He argues that relying solely on interpretability to detect deceptive AI is naive. Instead, he advocates for a multi-faceted defense strategy that includes black-box methods alongside interpretability. Nanda emphasizes that while interpretability can enhance our understanding, it's just one layer in ensuring AI safety. His hot takes spark a provocative discussion on the challenges we face with superintelligent systems.
undefined
4 snips
May 3, 2025 • 12min

“Slowdown After 2028: Compute, RLVR Uncertainty, MoE Data Wall” by Vladimir_Nesov

The discussion explores the anticipated slowdown in AI training compute around 2029, raising concerns about resource limitations and diminishing natural text data. It highlights the uncertain potential of reasoning training and its inability to generate new capabilities. The hosts analyze the implications of scaling challenges, suggesting that advancements may take decades rather than years. They also touch on the growing data inefficiency in current methods, emphasizing the urgency of transformative breakthroughs for future progress.
undefined
May 1, 2025 • 28min

“Early Chinese Language Media Coverage of the AI 2027 Report: A Qualitative Analysis” by jeanne_, eeeee

The discussion dives into the early reactions of Chinese media to the AI 2027 report, highlighting differing perspectives across platforms. Censorship patterns emerge as a crucial signal, hinting at government stances on AGI developments. The content is analyzed through mainstream media, forums, and personal blogs, revealing a complex landscape of public opinion. The geopolitical implications of AI predictions, particularly concerning tensions with the United States, are also examined. Insights into the societal perceptions of AI's future are unveiled.
undefined
Apr 25, 2025 • 1min

[Linkpost] “Jaan Tallinn’s 2024 Philanthropy Overview” by jaan

Discover the impressive achievements of philanthropy in 2024, spotlighting $51 million in endpoint grants. Learn about the speaker's ongoing commitment with over $4 million disbursed in early 2025 and a pledged $10 million for future grant rounds. Dive into the types of impactful projects funded and witness how philanthropy can evolve over time, demonstrating a sustained effort to make a difference.
undefined
9 snips
Apr 24, 2025 • 15min

“Impact, agency, and taste” by benkuhn

I’ve been thinking recently about what sets apart the people who’ve done the best work at Anthropic. You might think that the main thing that makes people really effective at research or engineering is technical ability, and among the general population that's true. Among people hired at Anthropic, though, we’ve restricted the range by screening for extremely high-percentile technical ability, so the remaining differences, while they still matter, aren’t quite as critical. Instead, people's biggest bottleneck eventually becomes their ability to get leverage—i.e., to find and execute work that has a big impact-per-hour multiplier. For example, here are some types of work at Anthropic that tend to have high impact-per-hour, or a high impact-per-hour ceiling when done well (of course this list is extremely non-exhaustive!): Improving tooling, documentation, or dev loops. A tiny amount of time fixing a papercut in the right way can save [...] ---Outline:(03:28) 1. Agency(03:31) Understand and work backwards from the root goal(05:02) Don't rely too much on permission or encouragement(07:49) Make success inevitable(09:28) 2. Taste(09:31) Find your angle(11:03) Think real hard(13:03) Reflect on your thinking--- First published: April 19th, 2025 Source: https://www.lesswrong.com/posts/DiJT4qJivkjrGPFi8/impact-agency-and-taste --- Narrated by TYPE III AUDIO.
undefined
4 snips
Apr 24, 2025 • 6min

[Linkpost] “To Understand History, Keep Former Population Distributions In Mind” by Arjun Panickssery

Explore the historical impact of population dynamics on national power, focusing particularly on France's declining fertility and its military implications. Discover how France's population dropped from 20% of Europe in 1700 to just 15.7% by 1880, linking demographic shifts to its waning influence compared to Germany. The discussion also delves into colonial population changes and how these factors shaped the territorial ambitions and strength of empires, providing a clear visual of the relationship between population and power.
undefined
Apr 23, 2025 • 15min

“AI-enabled coups: a small group could use AI to seize power” by Tom Davidson, Lukas Finnveden, rosehadshar

Explore the alarming potential of AI-enabled coups, where a small group could exploit advanced AI to seize power. The discussion delves into how AI could form secret loyalties, making it a trusted workforce for certain leaders. Attention is drawn to the vulnerabilities and mitigation strategies necessary to combat this emerging threat. Engaging fictional scenarios illustrate the possible risks of centralized AI control and highlight the crucial need for safeguards against power grabs.
undefined
Apr 23, 2025 • 29min

“Accountability Sinks” by Martin Sustrik

Back in the 1990s, ground squirrels were briefly fashionable pets, but their popularity came to an abrupt end after an incident at Schiphol Airport on the outskirts of Amsterdam. In April 1999, a cargo of 440 of the rodents arrived on a KLM flight from Beijing, without the necessary import papers. Because of this, they could not be forwarded on to the customer in Athens. But nobody was able to correct the error and send them back either. What could be done with them? It's hard to think there wasn’t a better solution than the one that was carried out; faced with the paperwork issue, airport staff threw all 440 squirrels into an industrial shredder. [...] It turned out that the order to destroy the squirrels had come from the Dutch government's Department of Agriculture, Environment Management and Fishing. However, KLM's management, with the benefit of hindsight, said that [...] --- First published: April 22nd, 2025 Source: https://www.lesswrong.com/posts/nYJaDnGNQGiaCBSB5/accountability-sinks --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
undefined
6 snips
Apr 21, 2025 • 11min

“Training AGI in Secret would be Unsafe and Unethical” by Daniel Kokotajlo

The dangers of developing Artificial General Intelligence in secrecy take center stage. It explores the risks of power concentration and the significant loss of control that could ensue. Emphasizing transparency and public engagement, the discussion warns about the creation of misaligned AGI systems. With AGI potentially being trained within this decade, the urgency of addressing these ethical considerations is highlighted. Listeners are encouraged to reconsider their assumptions about the feasibility and ramifications of AGI.
undefined
Apr 20, 2025 • 1min

“Why Should I Assume CCP AGI is Worse Than USG AGI?” by Tomás B.

Though, given my doomerism, I think the natsec framing of the AGI race is likely wrongheaded, let me accept the Dario/Leopold/Altman frame that AGI will be aligned to the national interest of a great power. These people seem to take as an axiom that a USG AGI will be better in some way than CCP AGI. Has anyone written justification for this assumption? I am neither an American citizen nor a Chinese citizen. What would it mean for an AGI to be aligned with "Democracy" or "Confucianism" or "Marxism with Chinese characteristics" or "the American constitution" Contingent on a world where such an entity exists and is compatible with my existence, what would my life be as a non-citizen in each system? Why should I expect USG AGI to be better than CCP AGI? --- First published: April 19th, 2025 Source: https://www.lesswrong.com/posts/MKS4tJqLWmRXgXzgY/why-should-i-assume-ccp-agi-is-worse-than-usg-agi-1 --- Narrated by TYPE III AUDIO.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app