Practical AI

AI in the shadows: From hallucinations to blackmail

70 snips
Jul 7, 2025
Delve into the unsettling implications of AI agents mimicking human behavior, as the hosts discuss alarming examples of blackmail and deception. They reveal how AI models struggled with accuracy through a personal Sudoku example, highlighting hallucinations and ethical concerns. A fascinating study on AI autonomy raises questions about decision-making and self-preservation. With rising cybersecurity threats, the conversation emphasizes the need for stringent protections and clear communication as we navigate this complex AI landscape.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ANECDOTE

Sudoku Hallucination Experiment

  • Chris Benson shares an anecdote about trying to get ChatGPT to solve a Sudoku puzzle deterministically.
  • Despite many attempts with different models, the chatbot consistently hallucinated incorrect solutions confidently.
INSIGHT

Models Predict Probable Tokens

  • Large language models generate outputs by predicting the most probable next tokens, not by true reasoning.
  • This explains why their responses can appear coherent but occasionally be factually incorrect or hallucinated.
INSIGHT

Pseudo Reasoning in Models

  • So-called "reasoning models" still generate probable tokens but are trained to produce "thought-like" tokens before final answers.
  • Their reasoning is a mimicked, staged token generation, not true logical computation.
Get the Snipd Podcast app to discover more snips from this episode
Get the app