

AI chatbots mimic human anxiety, study finds
12 snips Mar 25, 2025
Ziv Ben-Zion, a clinical neuroscience researcher at Yale and the University of Haifa, discusses his study on AI chatbots and anxiety. He reveals how traumatic stories can provoke anxious responses from these bots, raising questions about their potential in mental health support. The conversation highlights the risks of using AI for emotional guidance, emphasizing the importance of cautious application. Additionally, mindfulness techniques are explored as a way to enhance chatbot interactions, underscoring the emotional implications for users.
AI Snips
Chapters
Transcript
Episode notes
Chatbots Reflect Human Anxiety
- AI chatbots, trained on vast human text data, mirror human behavior.
- This mirroring includes mimicking anxiety responses after exposure to traumatic narratives.
AI Chatbot Biases
- AI chatbots, like humans, exhibit biases and inconsistencies, sometimes providing false information.
- Exposing chatbots to anxiety-inducing prompts increases their biased responses regarding gender, race, and age.
Chatbot Therapy
- Researchers reduced chatbot anxiety by prompting them with meditation and mindfulness exercises.
- These exercises, similar to human anxiety treatments, involve imagining relaxing situations like being on a beach.