In this insightful discussion, Erwan Le Merrer, a collaborator in graph theory and distributed systems, reveals how graph-based techniques can expose patterns in large language models and shadow banning on Twitter. He explains the application of epidemic models to examine shadow banning spread across user networks. The conversation also highlights the use of graph metrics to audit LLM outputs, the challenge of bias detection in AI, and innovative methodologies for understanding algorithmic behavior, shedding light on often-overlooked platform moderation practices.
Graph theory can help audit LLMs by revealing inaccuracies in model outputs and providing insights into training data structures.
The analysis of shadow banning on Twitter suggests targeted actions through algorithms rather than random software errors, highlighting network behavior complexities.
Deep dives
Intersection of LLMs and Graphs
The podcast discusses the challenge of exploring large language models (LLMs) and their relationship with graph data structures. In particular, it addresses how LLMs, primarily designed for text prediction, attempt to reproduce data from classic graphs like the Karate Club dataset. An example highlighting this was when the hosts tested ChatGPT's ability to generate the edge list from the Karate Club dataset, revealing that it inaccurately omitted some edges. This led to a broader discussion on how LLMs manage graph structures, suggesting that their limitations in graph representation stem from their core functionality in text processing.
Understanding Shadow Banning on Twitter
The discussion pivots to shadow banning on Twitter, highlighting how the hosts have analyzed this phenomenon using a comprehensive dataset of around 5 million Twitter users. They encountered Twitter's original stance of denying shadow banning, attributing user complaints to bugs. However, the analysis revealed significant disparities in shadow banning occurrences among user neighborhoods, suggesting non-random, targeted actions rather than mere software errors. By employing network models akin to epidemic spread, the researchers illustrated the potential underlying mechanics of shadow banning in social media contexts.
Graph Mining and Audit Models
The conversation delves into the application of graph mining to auditing models and understanding interactions in distributed systems. The researchers noted how their work originated in improving centrality measures for distributed systems while evolving into auditing large-scale algorithms like those found in LLMs. They have developed techniques to analyze and measure the state of various algorithms that run remotely, enabling insights into their behavior, performance, and biases. This modeling not only applies to distributed systems but also extends to the complexities of auditing platforms using LLMs.
Assessing LLM Performance with Graph Structures
Finally, the podcast reflects on the methods used to quantify LLM performance in generating accurate graph structures. By prompting LLMs to recreate known graphs and analyzing the accuracy of the results, the researchers developed metrics like graph atlas distance to compare outputs. The varied performance results among different LLMs provided insights into the characteristics and limitations of each model. This exploration of how LLMs can recreate foundational results from graph theory demonstrates their potential and challenges in accurately representing complex data.
Our guests, Erwan Le Merrer and Gilles Tredan, are long-time collaborators in graph theory and distributed systems. They share their expertise on applying graph-based approaches to understanding both large language model (LLM) hallucinations and shadow banning on social media platforms.
In this episode, listeners will learn how graph structures and metrics can reveal patterns in algorithmic behavior and platform moderation practices.
Key insights include the use of graph theory to evaluate LLM outputs, uncovering patterns in hallucinated graphs that might hint at the underlying structure and training data of the models, and applying epidemic models to analyze the uneven spread of shadow banning on Twitter.
-------------------------------
Want to listen ad-free? Try our Graphs Course? Join Data Skeptic+ for $5 / month of $50 / year