LessWrong (Curated & Popular) cover image

LessWrong (Curated & Popular)

“Connecting the Dots: LLMs can Infer & Verbalize Latent Structure from Training Data” by Johannes Treutlein, Owain_Evans

Jun 23, 2024
Researcher Johannes Treutlein and ML expert Owain Evans discuss LLMs' ability to infer latent information for tasks like defining functions and predicting city names without in-context learning. They showcase how LLMs can carry out tasks by leveraging training data without explicit reasoning.
17:56

Podcast summary created with Snipd AI

Quick takeaways

  • LLMs can infer latent information from training data for downstream tasks without in-context learning, demonstrating out-of-context reasoning capabilities.
  • Inductive out-of-context reasoning in LLMs raises AI safety concerns due to unmonitored acquisition of sensitive information and potential risks of deception.

Deep dives

Inductive Out-of-Context Reasoning in LLMs

LLMs can infer latent information from training data and utilize it for downstream tasks without in-context learning. Experimental results show that LLMs, fine-tuned on specific data like distances between cities, can deduce latent information such as the identity of unknown cities like Paris. Although effective in some cases, inductive out-of-context reasoning (OOCR) is shown to be unreliable, particularly with smaller LLMs tackling complex structures.

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner
Get the app