

The Geometry of Truth: Emergent Linear Structure in LLM Representation of True/False Datasets
Nov 30, 2023
In this podcast, Samuel Marks, a Postdoctoral Research Associate at Northeastern University, discusses his paper on the linear structure of true/false datasets in LLM representations. They explore how language models can linearly represent truth or falsehood, introduce a new probing technique called mass mean probing, and analyze the process of embedding truth in LLM models. They also discuss the future research directions and limitations of the paper.
Chapters
Transcript
Episode notes
1 2 3 4 5 6
Introduction
00:00 • 3min
Evaluation of GPT4's Capabilities and Hiring a Human Task Rabbit
02:44 • 2min
Analyzing Language Models' Representation of True and False Statements
04:20 • 23min
A New Probing Technique: Mass Mean Probing
27:08 • 3min
Embedding Truth in LLM and Adversarial Texts
30:30 • 7min
Future Research Directions and Limitations
37:31 • 3min