LessWrong (Curated & Popular)

LLMs for Alignment Research: a safety priority?

Apr 6, 2024
Gabriel Mukobi, author of a recent short story on LLMs, discusses prioritizing safety in AI research. They explore the role of programming and philosophy in safety work with LLMs, compare collaborative vs autonomous AI development, dive into AI hallucinations, data hunger in deep learning, and enhancing LLMs for safety through expert feedback.
Ask episode
Chapters
Transcript
Episode notes