Data Skeptic cover image

Data Skeptic

Latest episodes

undefined
Dec 7, 2018 • 20min

Data Ethics

 This week, Kyle interviews Scott Nestler on the topic of Data Ethics. Today, no ubiquitous, formal ethical protocol exists for data science, although some have been proposed. One example is the INFORMS Ethics Guidelines. Guidelines like this are rather informal compared to other professions, like the Hippocratic Oath. Yet not every profession requires such a formal commitment. In this episode, Scott shares his perspective on a variety of ethical questions specific to data and analytics.
undefined
Nov 30, 2018 • 34min

Escaping the Rabbit Hole

Kyle interviews Mick West, author of Escaping the Rabbit Hole: How to Debunk Conspiracy Theories Using Facts, Logic, and Respect about the nature of conspiracy theories, the people that believe them, and how to help people escape the belief in false information. Mick is also the creator of metabunk.org. The discussion explores conspiracies like chemtrails, 9/11 conspiracy theories, JFK assassination theories, and the flat Earth theory. We live in a complex world in which no person can have a sufficient understanding of all topics. It's only natural that some percentage of people will eventually adopt fringe beliefs. In this book, Mick provides a fantastic guide to helping individuals who have fallen into a rabbit hole of pseudo-science or fake news.
undefined
Nov 23, 2018 • 19min

[MINI] Theorem Provers

Fake news attempts to lead readers/listeners/viewers to conclusions that are not descriptions of reality.  They do this most often by presenting false premises, but sometimes by presenting flawed logic. An argument is only sound and valid if the conclusions are drawn directly from all the state premises, and if there exists a path of logical reasoning leading from those premises to the conclusion. While creating a theorem does feel to most mathematicians as a creative act of discovery, some theorems have been proven using nothing more than search.  All the "rules" of logic (like modus ponens) can be encoded into a computer program.  That program can start from the premises, applying various combinations of rules to inference new information, and check to see if the program has inference the desired conclusion or its negation.  This does seem like a mechanical process when painted in this light.  However, several challenges exist preventing any theorem prover from instantly solving all the open problems in mathematics.  In this episode, we discuss a bit about what those challenges are.  
undefined
Nov 16, 2018 • 32min

Automated Fact Checking

Fake news can be responded to with fact-checking. However, it's easier to create fake news than the fact check it. Full Fact is the UK's independent fact-checking organization. In this episode, Kyle interviews Mevan Babakar, head of automated fact-checking at Full Fact. Our discussion talks about the process and challenges in doing fact-checking. Full Fact has been exploring ways in which machine learning can assist in automating parts of the fact-checking process. Progress in areas like this allows journalists to be more effective and rapid in responding to new information.
undefined
Nov 9, 2018 • 30min

[MINI] Single Source of Truth

In mathematics, truth is universal.  In data, truth lies in the where clause of the query. As large organizations have grown to rely on their data more significantly for decision making, a common problem is not being able to agree on what the data is. As the volume and velocity of data grow, challenges emerge in answering questions with precision.  A simple question like "what was the revenue yesterday" could become mired in details.  Did your query account for transactions that haven't been finalized?  If I query again later, should I exclude orders that have been returned since the last query?  What time zone should I use?  The list goes on and on. In any large enough organization, you are also likely to find multiple copies if the same data.  Independent systems might record the same information with slight variance.  Sometimes systems will import data from other systems; a process which could become out of sync for several reasons. For any sufficiently large system, answering analytical questions with precision can become a non-trivial challenge.  The business intelligence community aspires to provide a "single source of truth" - one canonical place where data consumers can go to get precise, reliable, and trusted answers to their analytical questions.
undefined
Nov 2, 2018 • 45min

Detecting Fast Radio Bursts with Deep Learning

Fast radio bursts are an astrophysical phenomenon first observed in 2007. While many observations have been made, science has yet to explain the mechanism for these events. This has led some to ask: could it be a form of extra-terrestrial communication? Probably not. Kyle asks Gerry Zhang who works at the Berkeley SETI Research Center about this possibility and more importantly, about his applications of deep learning to detect fast radio bursts. Radio astronomy captures observations from space which can be converted to a waterfall chart or spectrogram. These data structures can be formatted in a visual way and also make great candidates for applying deep learning to the task of detecting the fast radio bursts.
undefined
Oct 26, 2018 • 25min

Being Bayesian

This episode explores the root concept of what it is to be Bayesian: describing knowledge of a system probabilistically, having an appropriate prior probability, know how to weigh new evidence, and following Bayes's rule to compute the revised distribution. We present this concept in a few different contexts but primarily focus on how our bird Yoshi sends signals about her food preferences. Like many animals, Yoshi is a complex creature whose preferences cannot easily be summarized by a straightforward utility function the way they might in a textbook reinforcement learning problem. Her preferences are sequential, conditional, and evolving. We may not always know what our bird is thinking, but we have some good indicators that give us clues.
undefined
Oct 19, 2018 • 33min

Modeling Fake News

This is our interview with Dorje Brody about his recent paper with David Meier, How to model fake news. This paper uses the tools of communication theory and a sub-topic called filtering theory to describe the mathematical basis for an information channel which can contain fake news.   Thanks to our sponsor Gartner.
undefined
Oct 12, 2018 • 27min

The Louvain Method for Community Detection

The podcast explores community detection in social networks using the Louvain Method. It discusses the concept of communities, the strength of connections within a community, and the theory behind the Louvain Method. The speakers also explore the potential use of the method in identifying interest-based communities and detecting fake news on social networks. Additionally, they discuss the spread of information within communities and the risk of spreading fake information.
undefined
Oct 5, 2018 • 32min

Cultural Cognition of Scientific Consensus

In this episode, our guest is Dan Kahan about his research into how people consume and interpret science news. In an era of fake news, motivated reasoning, and alternative facts, important questions need to be asked about how people understand new information. Dan is a member of the Cultural Cognition Project at Yale University, a group of scholars interested in studying how cultural values shape public risk perceptions and related policy beliefs. In a paper titled Cultural cognition of scientific consensus, Dan and co-authors Hank Jenkins‐Smith and Donald Braman discuss the "cultural cognition of risk" and establish experimentally that individuals tend to update their beliefs about scientific information through a context of their pre-existing cultural beliefs. In this way, topics such as climate change, nuclear power, and conceal-carry handgun permits often result in people. The findings of this and other studies tell us that on topics such as these, even when people are given proper information about a scientific consensus, individuals still interpret those results through the lens of their pre-existing cultural beliefs. The ‘cultural cognition of risk’ refers to the tendency of individuals to form risk perceptions that are congenial to their values. The study presents both correlational and experimental evidence confirming that cultural cognition shapes individuals’ beliefs about the existence of scientific consensus, and the process by which they form such beliefs, relating to climate change, the disposal of nuclear wastes, and the effect of permitting concealed possession of handguns. The implications of this dynamic for science communication and public policy‐making are discussed.

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode