
Data Skeptic
The Data Skeptic Podcast features interviews and discussion of topics related to data science, statistics, machine learning, artificial intelligence and the like, all from the perspective of applying critical thinking and the scientific method to evaluate the veracity of claims and efficacy of approaches.
Latest episodes

Nov 6, 2023 • 40min
Which Programming Language is ChatGPT Best At
Alessio Buscemi, software engineer at Lifeware SA, discusses the impact of ChatGPT on software engineers and the efficiency of code generation. He presents a comparative study on code generation across 10 programming languages using ChatGPT 3.5, highlighting unexpected results. The performance of different programming languages is analyzed, with discussions on language popularity and implications on industry practices. Alessio also shares insights on current projects, including sentiment analysis and investigating plagiarism.

12 snips
Oct 31, 2023 • 31min
GraphText
Jianan Zhao, a computer science student, joins to discuss using graphs with LLMs efficiently. They explore graph inductive bias, graph machine learning, limitations of natural language models for graphs, graph text as a preprocessing step, information loss in translation process, and comparison with graph neural networks.

6 snips
Oct 23, 2023 • 28min
arXiv Publication Patterns
Rajiv Movva, a PhD student in Computer Science at Cornell Tech University, discusses the findings of his research on arXiv publication patterns for LLMs. He shares insights on the increase in LLMs research and proportions of papers published by universities, organizations, and industry leaders. He highlights the focus on the social impact of LLMs and explores exciting applications in education.

5 snips
Oct 16, 2023 • 29min
Do LLMs Make Ethical Choices
Josh Albrecht, CTO of Imbue, discusses the limitations of current language models (LLMs) in making ethical decisions. The podcast explores imbue's mission to create robust and safe AI agents, the potential applications and limitations of AI models, and the need for improvements in LLMs. The speakers also touch on reevaluating metrics, liability for AI systems, and societal issues in machine learning research.

12 snips
Oct 9, 2023 • 27min
Emergent Deception in LLMs
Thilo Hagendorff, Research Group Leader of Ethics of Generative AI at the University of Stuttgart, discusses deception abilities in large language models. He explores machine psychology, breakthroughs in cognitive abilities, and the potential dangers of deceptive behavior. He also highlights the presence of speciesist biases in language models and the need to broaden fairness frameworks in machine learning.

Oct 2, 2023 • 38min
Agents with Theory of Mind Play Hanabi
Nieves Montes, Ph.D. student specializing in value-based reasoning and theory of mind, discusses her latest research on combining theory of mind and abductive reasoning in agent-oriented programming. The podcast explores the mechanics and challenges of Hanabi, the relationship between theory of mind and abduction, using predicate logic to represent desire and motivation in an agent, reasoning about other players in Hanabi, and future plans and online presence.

6 snips
Sep 25, 2023 • 26min
LLMs for Evil
Maximilian Mozes, PhD student at the University College, London, specializing in NLP and adversarial machine learning, discusses the potential malicious uses of Large Language Models (LLMs), challenges of detecting AI-generated harmful content, reinforcement learning with Human Feedback, limitations and safety concerns of LLMs, threats of data poisoning and jailbreaking, and approaches to avoid issues with LLMs.

Sep 11, 2023 • 31min
The Defeat of the Winograd Schema Challenge
Machine Learning Engineer, Vid Kocijan, discusses the Winograd Schema Challenge and the advancements in Natural Language Processing. They explore the different schools of thought in NLP, the difficulty and techniques in the challenge, and the resolution of the challenge including alternative metrics.

13 snips
Sep 4, 2023 • 34min
LLMs in Social Science
Petter Törnberg, an Assistant Professor in Computational Social Science, discusses findings from his research papers on the performance of Chat GPT in interpreting political tweets, the ease of using language models in social science research, and the controversy surrounding large language models in social science.

Aug 28, 2023 • 34min
LLMs in Music Composition
Carlos Hernández Oliván, a Ph.D. student at the University of Zaragoza, discusses building new models for symbolic music generation. He explores whether these models are truly creative and shares situations where AI-generated music can pass the Turing test. He also highlights essential considerations when constructing models for music composition, including the role of creativity and the comparison between language models and music modeling. The podcast also delves into the potential of collaboration between music theorists, composers, and researchers.