Data Skeptic cover image

Data Skeptic

Latest episodes

undefined
Oct 23, 2023 • 28min

arXiv Publication Patterns

Rajiv Movva, a PhD student in Computer Science at Cornell Tech University, discusses the findings of his research on arXiv publication patterns for LLMs. He shares insights on the increase in LLMs research and proportions of papers published by universities, organizations, and industry leaders. He highlights the focus on the social impact of LLMs and explores exciting applications in education.
undefined
Oct 16, 2023 • 29min

Do LLMs Make Ethical Choices

Josh Albrecht, CTO of Imbue, discusses the limitations of current language models (LLMs) in making ethical decisions. The podcast explores imbue's mission to create robust and safe AI agents, the potential applications and limitations of AI models, and the need for improvements in LLMs. The speakers also touch on reevaluating metrics, liability for AI systems, and societal issues in machine learning research.
undefined
Oct 9, 2023 • 27min

Emergent Deception in LLMs

Thilo Hagendorff, Research Group Leader of Ethics of Generative AI at the University of Stuttgart, discusses deception abilities in large language models. He explores machine psychology, breakthroughs in cognitive abilities, and the potential dangers of deceptive behavior. He also highlights the presence of speciesist biases in language models and the need to broaden fairness frameworks in machine learning.
undefined
Oct 2, 2023 • 38min

Agents with Theory of Mind Play Hanabi

Nieves Montes, Ph.D. student specializing in value-based reasoning and theory of mind, discusses her latest research on combining theory of mind and abductive reasoning in agent-oriented programming. The podcast explores the mechanics and challenges of Hanabi, the relationship between theory of mind and abduction, using predicate logic to represent desire and motivation in an agent, reasoning about other players in Hanabi, and future plans and online presence.
undefined
Sep 25, 2023 • 26min

LLMs for Evil

Maximilian Mozes, PhD student at the University College, London, specializing in NLP and adversarial machine learning, discusses the potential malicious uses of Large Language Models (LLMs), challenges of detecting AI-generated harmful content, reinforcement learning with Human Feedback, limitations and safety concerns of LLMs, threats of data poisoning and jailbreaking, and approaches to avoid issues with LLMs.
undefined
Sep 11, 2023 • 31min

The Defeat of the Winograd Schema Challenge

Machine Learning Engineer, Vid Kocijan, discusses the Winograd Schema Challenge and the advancements in Natural Language Processing. They explore the different schools of thought in NLP, the difficulty and techniques in the challenge, and the resolution of the challenge including alternative metrics.
undefined
Sep 4, 2023 • 34min

LLMs in Social Science

Petter Törnberg, an Assistant Professor in Computational Social Science, discusses findings from his research papers on the performance of Chat GPT in interpreting political tweets, the ease of using language models in social science research, and the controversy surrounding large language models in social science.
undefined
Aug 28, 2023 • 34min

LLMs in Music Composition

Carlos Hernández Oliván, a Ph.D. student at the University of Zaragoza, discusses building new models for symbolic music generation. He explores whether these models are truly creative and shares situations where AI-generated music can pass the Turing test. He also highlights essential considerations when constructing models for music composition, including the role of creativity and the comparison between language models and music modeling. The podcast also delves into the potential of collaboration between music theorists, composers, and researchers.
undefined
Aug 21, 2023 • 27min

Cuttlefish Model Tuning

Hongyi Wang, a Senior Researcher at Carnegie Mellon University, discusses his research paper on low-rank model training. He addresses the need for optimizing ML model training and the challenges of training large models. He introduces the Cuttlefish model, its use cases, and its superiority over the Low-Rank Adaptation technique. He also offers advice on entering the machine learning field.
undefined
Aug 15, 2023 • 39min

Which Professions Are Threatened by LLMs

On today’s episode, we have Daniel Rock, an Assistant Professor of Operations Information and Decisions at the Wharton School of the University of Pennsylvania. Daniel’s research focuses on the economics of AI and ML, specifically how digital technologies are changing the economy. Daniel discussed how AI has disrupted the job market in the past years. He also explained that it had created more winners than losers. Daniel spoke about the empirical study he and his coauthors did to quantify the threat LLMs pose to professionals. He shared how they used the O-NET dataset and the BLS occupational employment survey to measure the impact of LLMs on different professions. Using the radiology profession as an example, he listed tasks that LLMs could assume. Daniel broadly highlighted professions that are most and least exposed to LLMs proliferation. He also spoke about the risks of LLMs and his thoughts on implementing policies for regulating LLMs.

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode