

LLMs for Alignment Research: a safety priority?
Apr 6, 2024
Gabriel Mukobi, author of a recent short story on LLMs, discusses prioritizing safety in AI research. They explore the role of programming and philosophy in safety work with LLMs, compare collaborative vs autonomous AI development, dive into AI hallucinations, data hunger in deep learning, and enhancing LLMs for safety through expert feedback.
Chapters
Transcript
Episode notes
1 2 3 4 5
Introduction
00:00 • 3min
Exploring the Usefulness of Programming and Philosophy for Safety Research with Modern LLMs
02:31 • 4min
Collaborative AI Development vs. Autonomous Approach
06:16 • 8min
AI Hallucinations and Data Hunger in Deep Learning
14:03 • 5min
Enhancing Language Models for AI Safety through Expert Feedback
18:57 • 2min