LessWrong (30+ Karma) cover image

LessWrong (30+ Karma)

Latest episodes

undefined
Apr 4, 2025 • 17min

“LLM AGI will have memory, and memory changes alignment” by Seth Herd

Summary: When stateless LLMs are given memories they will accumulate new beliefs and behaviors, and that may allow their effective alignment to evolve. (Here "memory" is learning during deployment that is persistent beyond a single session.)[1] LLM agents will have memory: Humans who can't learn new things ("dense anterograde amnesia") are not highly employable for knowledge work. LLM agents that can learn during deployment seem poised to have a large economic advantage. Limited memory systems for agents already exist, so we should expect nontrivial memory abilities improving alongside other capabilities of LLM agents. Memory changes alignment: It is highly useful to have an agent that can solve novel problems and remember the solutions. Such memory includes useful skills and beliefs like "TPS reports should be filed in the folder ./Reports/TPS". They could also include learning skills for hiding their actions, and beliefs like "LLM agents are a type of [...] ---Outline:(01:26) Memory is useful for many tasks(05:11) Memory systems are ready for agentic use(09:00) Agents arent ready to direct memory systems(11:20) Learning new beliefs can functionally change goals and values(12:43) Value change phenomena in LLMs to date(14:27) Value crystallization and reflective stability as a result of memory(15:35) Provisional conclusions--- First published: April 4th, 2025 Source: https://www.lesswrong.com/posts/aKncW36ZdEnzxLo8A/llm-agi-will-have-memory-and-memory-changes-alignment --- Narrated by TYPE III AUDIO.
undefined
Apr 4, 2025 • 24min

“Will compute bottlenecks prevent a software intelligence explosion?” by Tom Davidson

Epistemic status – thrown together quickly. This is my best-guess, but could easily imagine changing my mind. Intro I recently copublished a report arguing that there might be a software intelligence explosion (SIE) – once AI R&D is automated (i.e. automating OAI), the feedback loop of AI improving AI algorithms could accelerate more and more without needing more hardware. If there is an SIE, the consequences would obviously be massive. You could shoot from human-level to superintelligent AI in a few months or years; by default society wouldn’t have time to prepare for the many severe challenges that could emerge (AI takeover, AI-enabled human coups, societal disruption, dangerous new technologies, etc). The best objection to an SIE is that progress might be bottlenecked by compute. We discuss this in the report, but I want to go into much more depth because it's a powerful objection [...] ---Outline:(00:19) Intro(01:47) The compute bottleneck objection(01:51) Intuitive version(02:58) Economist version(09:13) Counterarguments to the compute bottleneck objection(20:11) Taking stock--- First published: April 4th, 2025 Source: https://www.lesswrong.com/posts/XDF6ovePBJf6hsxGj/will-compute-bottlenecks-prevent-a-software-intelligence-1 --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
undefined
Apr 4, 2025 • 1h 28min

“AI #110: Of Course You Know…” by Zvi

Yeah. That happened yesterday. This is real life. I know we have to ensure no one notices Gemini 2.5 Pro, but this is rediculous. That's what I get for trying to go on vacation to Costa Rica, I suppose. I debated waiting for the market to open to learn more. But f*** it, we ball. Table of Contents Also this week: More Fun With GPT-4o Image Generation, OpenAI #12: Battle of the Board Redux and Gemini 2.5 Pro is the New SoTA. The New Tariffs Are How America Loses. This is somehow real life. Is AI Now Impacting the Global Economy Bigly? Asking the wrong questions. Language Models Offer Mundane Utility. Is it good enough for your inbox yet? Language Models Don’t Offer Mundane Utility. Why learn when you can vibe? Huh, Upgrades. GPT-4o, Gemini 2.5 Pro, and [...] ---Outline:(00:35) The New Tariffs Are How America Loses(07:35) Is AI Now Impacting the Global Economy Bigly?(12:07) Language Models Offer Mundane Utility(14:28) Language Models Don't Offer Mundane Utility(15:09) Huh, Upgrades(17:09) On Your Marks(23:27) Choose Your Fighter(25:51) Jevons Paradox Strikes Again(26:25) Deepfaketown and Botpocalypse Soon(31:47) They Took Our Jobs(33:02) Get Involved(33:41) Introducing(35:25) In Other AI News(37:17) Show Me the Money(43:12) Quiet Speculations(47:24) The Quest for Sane Regulations(53:52) Don't Maim Me Bro(57:29) The Week in Audio(57:54) Rhetorical Innovation(01:03:39) Expect the Unexpected(01:05:48) Open Weights Are Unsafe and Nothing Can Fix This(01:14:09) Anthropic Modifies Its Responsible Scaling Policy(01:18:04) If You're Not Going to Take This Seriously(01:20:24) Aligning a Smarter Than Human Intelligence is Difficult(01:23:54) Trust the Process(01:26:30) People Are Worried About AI Killing Everyone(01:26:52) The Lighter Side--- First published: April 3rd, 2025 Source: https://www.lesswrong.com/posts/bc8DQGvW3wiAWYibC/ai-110-of-course-you-know --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
undefined
Apr 3, 2025 • 11min

“The Rise of Hyperpalatability” by Jack

Hello, this is my first post here. I was told by a friend that I should post here. This is from a series of works that I wrote with strict structural requirements. I have performed minor edits to make the essay more palatable for human consumption. This work is an empirical essay on a cycle of hunger to satiatiation to hyperpalatability that I have seen manifested in multiple domains ranging from food to human connection. My hope is that you will gain some measure of appreciation for how we have shifted from a society geared towards sufficent production to one based on significant curation. Hyperpalatable Food For the majority of human history we lived in a production market for food. We searched for that which tasted well but there was never enough to fill the void. Only the truly elite could afford to import enough food [...] ---Outline:(00:44) Hyperpalatable Food(03:19) Hyperpalatable Media(05:40) Hyperpalatable Connection(08:08) Hyperpalatable Systems--- First published: April 2nd, 2025 Source: https://www.lesswrong.com/posts/bLTjZbCBanpQ9Kxgs/the-rise-of-hyperpalatability --- Narrated by TYPE III AUDIO.
undefined
Apr 3, 2025 • 9min

“Why Have Sentence Lengths Decreased?” by Arjun Panickssery

“In the loveliest town of all, where the houses were white and high and the elms trees were green and higher than the houses, where the front yards were wide and pleasant and the back yards were bushy and worth finding out about, where the streets sloped down to the stream and the stream flowed quietly under the bridge, where the lawns ended in orchards and the orchards ended in fields and the fields ended in pastures and the pastures climbed the hill and disappeared over the top toward the wonderful wide sky, in this loveliest of all towns Stuart stopped to get a drink of sarsaparilla.” — 107-word sentence from Stuart Little (1945) Sentence lengths have declined. The average sentence length was 49 for Chaucer (died 1400), 50 for Spenser (died 1599), 42 for Austen (died 1817), 20 for Dickens (died 1870), 21 for Emerson (died 1882), 14 [...] --- First published: April 3rd, 2025 Source: https://www.lesswrong.com/posts/xYn3CKir4bTMzY5eb/why-have-sentence-lengths-decreased --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
undefined
Apr 3, 2025 • 3min

“Announcing ILIAD2: ODYSSEY” by Alexander Gietelink Oldenziel, windows

We are pleased to announce ILIAD2: ODYSSEY—a 5-day conference bringing together 100+ researchers to build scientific foundations for AI alignment. This is the 2nd iteration of ILIAD, which was first held in the summer of 2024. ***Apply to attend by June 1!*** When: Aug 25-29, 2025 Where: @Lighthaven (Berkeley, US) What: Unconference with participant-led programming, 100+ attendees. Who: Currently confirmed attendees include Scott Garrabrant, Daniel Murfet, James Crutchfield, Jesse Hoogland, and Adam Shai Costs: Tickets are free. Financial support for travel and accommodations is available on a needs basis.  See our website here. For any questions, email iliadconference@gmail.com About ODYSSEY ODYSSEY is a 100+ person conference about alignment with a mathematical focus. ODYSSEY will feature an unconference format—meaning that participants can propose and lead their own sessions. We believe that this is the best way to release the latent creative energies in everyone attending. The [...] ---Outline:(00:28) \*\*\*Apply to attend by June 1!\*\*\*(01:26) About ODYSSEY(02:10) Financial Support(02:23) Proceedings(02:51) Artwork--- First published: April 3rd, 2025 Source: https://www.lesswrong.com/posts/WP7TbzzM39agMS77e/announcing-iliad2-odyssey-1 --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
undefined
Apr 3, 2025 • 55min

“AI 2027: What Superintelligence Looks Like” by Daniel Kokotajlo, Thomas Larsen, elifland, Scott Alexander, Jonas V, romeo

In 2021 I wrote what became my most popular blog post: What 2026 Looks Like. I intended to keep writing predictions all the way to AGI and beyond, but chickened out and just published up till 2026. Well, it's finally time. I'm back, and this time I have a team with me: the AI Futures Project. We've written a concrete scenario of what we think the future of AI will look like. We are highly uncertain, of course, but we hope this story will rhyme with reality enough to help us all prepare for what's ahead. You really should go read it on the website instead of here, it's much better. There's a sliding dashboard that updates the stats as you scroll through the scenario! But I've nevertheless copied the first half of the story below. I look forward to reading your comments.Mid 2025: Stumbling Agents The [...] ---Outline:(01:35) Mid 2025: Stumbling Agents(03:13) Late 2025: The World's Most Expensive AI(08:34) Early 2026: Coding Automation(10:49) Mid 2026: China Wakes Up(13:48) Late 2026: AI Takes Some Jobs(15:35) January 2027: Agent-2 Never Finishes Learning(18:20) February 2027: China Steals Agent-2(21:12) March 2027: Algorithmic Breakthroughs(23:58) April 2027: Alignment for Agent-3(27:26) May 2027: National Security(29:50) June 2027: Self-improving AI(31:36) July 2027: The Cheap Remote Worker(34:35) August 2027: The Geopolitics of Superintelligence(40:43) September 2027: Agent-4, the Superhuman AI Researcher--- First published: April 3rd, 2025 Source: https://www.lesswrong.com/posts/TpSFoqoG2M5MAAesg/ai-2027-what-superintelligence-looks-like-1 --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
undefined
Apr 3, 2025 • 19min

“More Fun With GPT-4o Image Generation” by Zvi

Greetings from Costa Rica! The image fun continues. We Are Going to Need A Bigger Compute Budget Fun is being had by all, now that OpenAI has dropped its rule about not mimicking existing art styles. Sam Altman (2:11pm, March 31): the chatgpt launch 26 months ago was one of the craziest viral moments i’d ever seen, and we added one million users in five days. We added one million users in the last hour. Sam Altman (8:33pm, March 31): chatgpt image gen now rolled out to all free users! Slow down. We’re going to need you to have a little less fun, guys. Sam Altman: it's super fun seeing people love images in chatgpt. but our GPUs are melting. we are going to temporarily introduce some rate limits while we work on making it more efficient. hopefully won’t be [...] ---Outline:(00:15) We Are Going to Need A Bigger Compute Budget(02:21) Defund the Fun Police(06:06) Fun the Artists(12:22) The No Fun Zone(14:49) So Many Other Things to Do(15:08) Self Portrait--- First published: April 3rd, 2025 Source: https://www.lesswrong.com/posts/GgNdBz5FhvqMJs5Qv/more-fun-with-gpt-4o-image-generation --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
undefined
Apr 3, 2025 • 7min

“How To Believe False Things” by Eneasz

Intro [you can skip this section if you don’t need context and just want to know how I could believe such a crazy thing] In my chat community: “Open Play” dropped, a book that says there's no physical difference between men and women so there shouldn’t be separate sports leagues. Boston Globe says their argument is compelling. Discourse happens, which is mostly a bunch of people saying “lololololol great trolling, what idiot believes such obvious nonsense?” I urge my friends to be compassionate to those sharing this. Because “until I was 38 I thought Men's World Cup team vs Women's World Cup team would be a fair match and couldn't figure out why they didn't just play each other to resolve the big pay dispute.” This is the one-line summary of a recent personal world-shattering I describe in a lot more detail here (link). I’ve had multiple people express [...] --- First published: April 2nd, 2025 Source: https://www.lesswrong.com/posts/fLm3F7vequjr8AhAh/how-to-believe-false-things --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
undefined
Apr 2, 2025 • 22min

“Show, not tell: GPT-4o is more opinionated in images than in text” by Daniel Tan, eggsyntax

Epistemic status: This should be considered an interim research note. Feedback is appreciated. Introduction We increasingly expect language models to be ‘omni-modal’, i.e. capable of flexibly switching between images, text, and other modalities in their inputs and outputs. In order to get a holistic picture of LLM behaviour, black-box LLM psychology should take into account these other modalities as well. In this project, we do some initial exploration of image generation as a modality for frontier model evaluations, using GPT-4o's image generation API. GPT-4o is one of the first LLMs to produce images natively rather than creating a text prompt which is sent to a separate image model, outputting images and autoregressive token sequences (ie in the same way as text). We find that GPT-4o tends to respond in a consistent manner to similar prompts. We also find that it tends to more readily express emotions [...] ---Outline:(00:53) Introduction(02:19) What we did(03:47) Overview of results(03:54) Models more readily express emotions / preferences in images than in text(05:38) Quantitative results(06:25) What might be going on here?(08:01) Conclusions(09:04) Acknowledgements(09:16) Appendix(09:28) Resisting their goals being changed(09:51) Models rarely say they'd resist changes to their goals(10:14) Models often draw themselves as resisting changes to their goals(11:31) Models also resist changes to specific goals(13:04) Telling them 'the goal is wrong' mitigates this somewhat(13:43) Resisting being shut down(14:02) Models rarely say they'd be upset about being shut down(14:48) Models often depict themselves as being upset about being shut down(17:06) Comparison to other topics(17:10) When asked about their goals being changed, models often create images with negative valence(17:48) When asked about different topics, models often create images with positive valence(18:56) Other exploratory analysis(19:09) Sandbagging(19:31) Alignment faking(19:55) Negative reproduction results(20:23) On the future of humanity after AGI(20:50) On OpenAI's censorship and filtering(21:15) On GPT-4o's lived experience:--- First published: April 2nd, 2025 Source: https://www.lesswrong.com/posts/XgSYgpngNffL9eC8b/show-not-tell-gpt-4o-is-more-opinionated-in-images-than-in --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode