AI Safety Fundamentals: Alignment cover image

Is Power-Seeking AI an Existential Risk?

AI Safety Fundamentals: Alignment

00:00

Exploring Robust Forms of Practical P.S. Alignment and Controlling Capabilities

This chapter delves into the concept of robust practical P.S. alignment in AI systems, discussing the search for problematic objectives, reading objectives from internal states, and the potential of short-term objectives in mitigating power-seeking behaviors. It also examines the challenges and difficulties in ensuring myopia with certain search techniques and long-term training processes.

Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner