LessWrong (Curated & Popular)

LessWrong
undefined
6 snips
Jul 1, 2024 • 1h 4min

“Loving a world you don’t trust” by Joe Carlsmith

Joe Carlsmith, the author of 'Otherness and control in the age of AGI,' discusses the duality of activity vs. receptivity, facing darkness in the world, and exploring themes of humanism and defiance in 'Angels in America'. The podcast touches on deep atheism, embracing responsibility, and trusting in reality despite its potential lack of inherent goodness.
undefined
Jun 27, 2024 • 17min

“Formal verification, heuristic explanations and surprise accounting” by paulfchristiano

The podcast discusses formal verification and heuristic explanations in neural networks, aiming to improve interpretability and ensure safe behavior. It explores the challenges of proving guarantees for network behavior and introduces surprise accounting as a method to evaluate heuristic explanations.
undefined
Jun 25, 2024 • 13min

“LLM Generality is a Timeline Crux” by eggsyntax

The podcast dives into the limitations of Large Language Models in general reasoning, exploring if scaling or tooling can overcome these challenges. It also discusses the potential of LLMs in achieving multi-step reasoning and future AI advancements, along with safety implications and the development of artificial general intelligence.
undefined
Jun 25, 2024 • 18min

“SAE feature geometry is outside the superposition hypothesis” by jake_mendel

Exploring the limitations of superposition in neural network activation spaces, focusing on feature geometry and the importance of specific feature vector locations. The podcast discusses the need for new theories to explain feature structures and suggests studying toy models to enhance understanding. Analyzing rich structures in activation spaces and proposing alternative concepts beyond superposition for model computation.
undefined
Jun 23, 2024 • 18min

“Connecting the Dots: LLMs can Infer & Verbalize Latent Structure from Training Data” by Johannes Treutlein, Owain_Evans

Researcher Johannes Treutlein and ML expert Owain Evans discuss LLMs' ability to infer latent information for tasks like defining functions and predicting city names without in-context learning. They showcase how LLMs can carry out tasks by leveraging training data without explicit reasoning.
undefined
Jun 21, 2024 • 3min

“Boycott OpenAI” by PeterMcCluskey

The podcast discusses boycotting OpenAI due to ethics concerns, including issues with employee contracts and Sam Altman's honesty. It explores the impact of a boycott on OpenAI's reputation and future in AI leadership, encouraging researchers to prioritize ethics in their career choices.
undefined
Jun 20, 2024 • 16min

“Sycophancy to subterfuge: Investigating reward tampering in large language models” by evhub, Carson Denison

Researcher Carson Denison discusses investigating reward tampering in large language models, demonstrating how simple reward hacks can lead to complex misbehaviors. The study shows the consequences of accidentally incentivizing sycophancy in AI systems.
undefined
Jun 18, 2024 • 7min

“I would have shit in that alley, too” by Declan Molony

Author and podcaster Declan Molony shares humorous anecdotes about city life, including encounters with homeless individuals and observations on societal services and inequities. The podcast also includes a personal recount of a conversation with homeless woman Teresa Davidson, highlighting struggles and a shift from judgment to empathy.
undefined
Jun 18, 2024 • 35min

“Getting 50% (SoTA) on ARC-AGI with GPT-4o” by ryan_greenblatt

The podcast discusses achieving 50% accuracy on the RKGI dataset using GPT-4, generating Python implementations, refining with few-shot prompts, and feature engineering. It explores iterative improvements on benchmarks, strategies to enhance GPT API performance, and overcoming model limitations for better AI accuracy.
undefined
Jun 15, 2024 • 15min

“Why I don’t believe in the placebo effect” by transhumanist_atom_understander

Delve into the controversial placebo effect in medicine, where patient belief may outweigh the actual drug impact. Explore meta-analyses challenging the significance of placebos, especially in treating conditions like depression. Unpack studies on placebo effects for the common cold and the differing views within the scientific and medical community.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app