AXRP - the AI X-risk Research Podcast cover image

32 - Understanding Agency with Jan Kulveit

AXRP - the AI X-risk Research Podcast

00:00

Understanding Self-Awareness in Language Models

This chapter explores the conceptual frameworks of large language models, emphasizing active inference and its contrast with reinforcement learning. The speakers discuss how feedback loops can enhance an AI's self-awareness, illustrating this with analogies to human navigation experiences. They also delve into the complexities of fine-tuning and the implications of behavioral changes in language models, highlighting the philosophical and societal considerations of AI development.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app