LessWrong (Curated & Popular) cover image

[HUMAN VOICE] "A case for AI alignment being difficult" by jessicata

LessWrong (Curated & Popular)

00:00

Challenges of AI Alignment and Consequentialism

This chapter explores the challenges of aligning AI systems with human values and the concept of consequentialism in problem-solving. It discusses the limitations of short-term and long-term consequentialism, the potential misalignment between AI systems and human values, and the difficulties in specifying courage-ability and low impact. The chapter also highlights the trade-offs and concerns related to current machine learning paradigms and their potential for misaligned long-term agency.

Play episode from 06:16
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app