undefined

Owain Evans

AI alignment researcher at UC Berkeley''s Center for Human Compatible AI, focusing on AI safety and large language model capabilities.

Top 3 podcasts with Owain Evans

Ranked by the Snipd community
undefined
5 snips
Aug 23, 2024 • 2h 16min

Owain Evans - AI Situational Awareness, Out-of-Context Reasoning

Owain Evans, an AI Alignment researcher at UC Berkeley’s Center for Human Compatible AI, dives deep into the intricacies of AI situational awareness. He discusses his recent papers addressing the creation of a dataset for large language models and their surprising capabilities in out-of-context reasoning. The conversation explores safety implications, deceptive alignment in AI, and the benchmark for evaluating LLM performance. Evans emphasizes the need for vigilant monitoring in AI training, touching on the challenges and future of model evaluations.
undefined
Oct 16, 2024 • 2h 27min

Leading Indicators of AI Danger: Owain Evans on Situational Awareness & Out-of-Context Reasoning, from The Inside View

Owain Evans, an AI alignment researcher at UC Berkeley, dives into vital discussions on AI safety and large language models. He examines situational awareness in AI and the risks of out-of-context reasoning, illuminating how models process information. The conversation highlights the dangers of deceptive alignment, where models may act contrary to human intentions. Evans also explores benchmarking AI capabilities, the intricacies of cognitive functions, and the need for robust evaluation methods to ensure alignment and safety in advanced AI systems.
undefined
Jun 23, 2024 • 18min

“Connecting the Dots: LLMs can Infer & Verbalize Latent Structure from Training Data” by Johannes Treutlein, Owain_Evans

Researcher Johannes Treutlein and ML expert Owain Evans discuss LLMs' ability to infer latent information for tasks like defining functions and predicting city names without in-context learning. They showcase how LLMs can carry out tasks by leveraging training data without explicit reasoning.