LessWrong (Curated & Popular)

LessWrong
undefined
Jan 9, 2025 • 25min

“What Indicators Should We Watch to Disambiguate AGI Timelines?” by snewman

The podcast dives into the murky waters of AGI timelines, capturing the contrasting theories on whether we're on the brink of a technological revolution or simply beginning a journey. It dissects the slow versus fast scenarios of AGI development, highlighting the hurdles current AI faces. The discussion also explores crucial indicators to watch for signs of a rapidly approaching AGI. With ongoing advancements, the narrative raises both excitement and skepticism about what lies ahead in the world of artificial intelligence.
undefined
Jan 8, 2025 • 1h 19min

“How will we update about scheming?” by ryan_greenblatt

I mostly work on risks from scheming (that is, misaligned, power-seeking AIs that plot against their creators such as by faking alignment). Recently, I (and co-authors) released "Alignment Faking in Large Language Models", which provides empirical evidence for some components of the scheming threat model.One question that's really important is how likely scheming is. But it's also really important to know how much we expect this uncertainty to be resolved by various key points in the future. I think it's about 25% likely that the first AIs capable of obsoleting top human experts[1] are scheming. It's really important for me to know whether I expect to make basically no updates to my P(scheming)[2] between here and the advent of potentially dangerously scheming models, or whether I expect to be basically totally confident one way or another by that point (in the same way that, though I might [...] ---Outline:(03:20) My main qualitative takeaways(04:56) Its reasonably likely (55%), conditional on scheming being a big problem, that we will get smoking guns.(05:38) Its reasonably likely (45%), conditional on scheming being a big problem, that we wont get smoking guns prior to very powerful AI.(15:59) My P(scheming) is strongly affected by future directions in model architecture and how the models are trained(16:33) The model(22:38) Properties of the AI system and training process(23:02) Opaque goal-directed reasoning ability(29:24) Architectural opaque recurrence and depth(34:14) Where do capabilities come from?(39:42) Overall distribution from just properties of the AI system and training(41:20) Direct observations(41:43) Baseline negative updates(44:35) Model organisms(48:21) Catching various types of problematic behavior(51:22) Other observations and countermeasures(52:02) Training processes with varying (apparent) situational awareness(54:05) Training AIs to seem highly corrigible and (mostly) myopic(55:46) Reward hacking(57:28) P(scheming) under various scenarios (putting aside mitigations)(01:05:19) An optimistic and a pessimistic scenario for properties(01:10:26) Conclusion(01:11:58) Appendix: Caveats and definitions(01:14:49) Appendix: Capabilities from intelligent learning algorithmsThe original text contained 15 footnotes which were omitted from this narration. --- First published: January 6th, 2025 Source: https://www.lesswrong.com/posts/aEguDPoCzt3287CCD/how-will-we-update-about-scheming --- Narrated by TYPE III AUDIO.
undefined
7 snips
Jan 8, 2025 • 20min

“OpenAI #10: Reflections” by Zvi

Sam Altman, CEO of OpenAI, discusses the complex landscape of AI leadership. He shares personal reflections on governance challenges following significant board disputes. Altman highlights the potential for advanced AI to transform productivity by 2025, while expressing concerns about the ethical implications of this technology. He also addresses accusations of fake news surrounding OpenAI and emphasizes the importance of trust and diverse perspectives in navigating the future of artificial intelligence.
undefined
8 snips
Jan 7, 2025 • 2min

“Maximizing Communication, not Traffic” by jefftk

Explore the idea of prioritizing clear communication in writing over chasing traffic. The discussion critiques clickbait tactics popular among content creators and stresses the value of genuine idea sharing. Discover the unique approach of including full posts in RSS feeds, emphasizing understanding rather than mere clicks. This perspective encourages writers to focus on the essence of their messages rather than the metrics that drive ad revenue.
undefined
Jan 2, 2025 • 44min

“What’s the short timeline plan?” by Marius Hobbhahn

Marius Hobbhahn, an insightful author and thinker on AI risks, dives into the urgent necessity for a concrete short timeline plan in AI development. He argues that with advancements possibly leading to AGI by 2027, detailed strategies are crucial for safety and effectiveness. Hobbhahn stresses the importance of collaboration among researchers and advocates for improved interpretability and monitoring methods. He elaborates on the need for a layered approach to control and evaluation, ensuring robust safety measures as technology rapidly evolves.
undefined
Dec 30, 2024 • 1h 57min

“Shallow review of technical AI safety, 2024” by technicalities, Stag, Stephen McAleese, jordine, Dr. David Mathers

Dive into the crucial realm of technical AI safety with engaging discussions on current research agendas and the complexities of AI alignment. Discover the challenges researchers face as they strive for responsible AI development. The conversation touches on interpretability, control measures, and the importance of goal robustness. Uncover innovative safety designs and the role of collaborative efforts in mitigating existential risks. This insightful overview is perfect for anyone curious about navigating the evolving landscape of AI safety.
undefined
13 snips
Dec 29, 2024 • 29min

“By default, capital will matter more than ever after AGI” by L Rudolf L

L. Rudolf L, an insightful author known for discussing the implications of AGI on society, dives deep into the future of capital in a world with labor-replacing AI. He argues that capital will become more essential than ever, transforming power dynamics and potentially leading to increased inequality. Rudolf challenges the notion that money will lose its value, exploring the risks of a divided society. He also examines Universal Basic Income as a response to these changes, questioning the evolving relationship between citizens and their governments.
undefined
Dec 28, 2024 • 39min

“Review: Planecrash” by L Rudolf L

Dive into a bizarre blend of fantasy and logic where math lectures sneak into a whimsical narrative. Discover characters navigating the chaotic realm of Dath-Elon, grappling with competence and decision theory. Explore philosophical musings on rationality that question conventional thinking and governance challenges. The intriguing dynamic between superintelligent beings and human emotions unfolds, leading to quirky tangents and deep reflections. It’s a wild journey that fuses intellectual rigor with storytelling.
undefined
Dec 26, 2024 • 14min

“The Field of AI Alignment: A Postmortem, and What To Do About It” by johnswentworth

johnswentworth, an insightful author from LessWrong, dissects the current state of AI alignment research. He uses an engaging metaphor about searching for keys under a streetlight to illustrate researchers' focus on easier problems while neglecting existential threats. The conversation shifts towards the urgent need for a recruitment overhaul, advocating for advanced technical skills to foster innovative approaches. Overall, the dialogue challenges existing paradigms and emphasizes tackling the real challenges in AI safety.
undefined
5 snips
Dec 23, 2024 • 11min

“When Is Insurance Worth It?” by kqr

In this insightful discussion, the complexities of insurance are unraveled. Misunderstandings about the necessity and value of insurance are addressed, debunking common myths. The podcast emphasizes using the Kelly Insurance Calculator to make informed decisions based on mathematical reasoning. Specific scenarios like motorcycle insurance and the impact of deductibles are explored. Ultimately, it highlights how understanding probabilities can safeguard long-term financial health.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app