
LessWrong (Curated & Popular) “the case for CoT unfaithfulness is overstated” by nostalgebraist
Sep 30, 2024
The conversation dives into the skepticism surrounding chain-of-thought (CoT) explanations from large language models. It challenges the notion that these explanations are entirely untrustworthy. Listeners are encouraged to reconsider the insights CoTs can provide, despite their flaws. The discussion emphasizes recognizing the unique benefits of CoTs compared to other reasoning methods. A call to explore what we can learn from model reasoning is made, urging a more nuanced view of model-generated reasoning processes.
Chapters
Transcript
Episode notes
