The Gradient: Perspectives on AI cover image

Laura Weidinger: Ethical Risks, Harms, and Alignment of Large Language Models

The Gradient: Perspectives on AI

00:00

AI Ethics - You Can't Explain Deep Learning?

I think it's probably a safe assumption that these policy research teams are looking at this AI ethics type of work. And on the other side, you also mentioned part of the goal is how to mitigate the harms and that we don't have enough mechanisms for that. The paper was really driven by the technical safety team at DeepMind. So there is a team that is looking at reward, misspecification, for example, technical misalignment. We got involved from the ethics point of view as well.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app