Episode 36: About That 'Dangerous Capabilities' Fanfiction (feat. Ali Alkhatib), June 24 2024
Jul 19, 2024
auto_awesome
Computer scientist Ali Alkhatib discusses Google DeepMind's flawed study on dangerous capabilities of large language models, emphasizing the social implications of AI. Critiques on poorly developed AI models, misinterpretations in AI research, hacker tools extracting data, and the limitations of tech news solutions are also explored. The conversation challenges hierarchical thinking in AI benchmarks and deceptive rhetoric in promoting harmful ideas.
Big tech companies using preprint servers to avoid peer review, influence public AI conversation.
Anthropic's Cloud 3.5 Sonnet benchmarks criticized for hierarchical implications.
AI technology used to moderate emotional expressions in customer service sparks ethical debate.
Deep dives
Cloud 3.5 Sonnet Benchmarks Segmented by Educational Stages Spark Controversy
In a post from anthropic on LinkedIn, the Cloud 3.5 Sonnet benchmarks are revealed to have segments like 'graduate-level reasoning,' 'undergraduate-level knowledge,' and 'grade school math,' drawing criticism for the hierarchical implications embedded in the model names.
New York City Audit of ShotSpotter Technology Reveals 87% False Alarms in Responses to Gunfire
An audit in New York City uncovers that ShotSpotter technology triggers false alarms 87% of the time when responding to gunfire, highlighting the inefficacy of the system.
AI Designed to Modify Customer Anger in Phone Calls Sparks Controversy
Softbank's project focused on using AI to soften customer phone calls by altering the tonality of enraged customer complaints prompts debate on the impacts of using technology to moderate emotional expressions.
Microsoft's Recall AI Tool Criticized for Lack of Data Security
A Wired article reports on a hacker tool that extracts data collected by Microsoft's Recall AI, raising concerns about data privacy and security breaches arising from the AI's screen recording capabilities.
Criticism Surrounds Softbank's Initiative to Mitigate Customer Anger Using AI in Phone Calls
Softbank's venture to create AI technology capable of altering the anger in customer phone calls stirs debate on the ethical implications of using AI to manage emotional interactions in customer service.
Misleading Claims on AI's Role in Softening Customer Anger Resurface Debates
Concerns arise over the ethical and practical implications of Softbank's project aiming to moderate customer anger in phone calls through AI tools, sparking discussions on the appropriateness of such interventions in customer service interactions.
When is a research paper not a research paper? When a big tech company uses a preprint server as a means to dodge peer review -- in this case, of their wild speculations on the 'dangerous capabilities' of large language models. Ali Alkhatib joins Emily to explain why a recent Google DeepMind document about the hunt for evidence that LLMs might intentionally deceive us was bad science, and yet is still influencing the public conversation about AI.
Ali Alkhatib is a computer scientist and former director of the University of San Francisco’s Center for Applied Data Ethics. His research focuses on human-computer interaction, and why our technological problems are really social – and why we should apply social science lenses to data work, algorithmic justice, and even the errors and reality distortions inherent in AI models.