

Protecting us with the Database of Evil
Nov 16, 2022
Matar Haller, VP of Data at ActiveFence, leads efforts to combat online harms like disinformation and extremism using AI and expert insights. He shares the importance of adapting technologies for ongoing challenges, especially in online safety for children and effective content moderation. The conversation highlights how robust safety measures can enhance user trust and discusses the intricate balance of using data-driven models to identify harmful content. Matar also touches on the emotional challenges of dealing with distressing material in their mission-driven work.
AI Snips
Chapters
Transcript
Episode notes
Prioritize Online Safety
- Prioritize online safety from the start, as it's now a competitive advantage and basic user expectation.
- Integrate trust and safety measures early in your platform's development, not as an afterthought.
Limitations of Keyword Moderation
- Traditional content moderation methods, like keyword blocking, are insufficient due to evolving adversarial tactics.
- Contextual understanding of language is crucial, as simple keyword-based moderation can't address nuances like sarcasm or quoted slurs.
Spectrum of Online Harms
- Online harms exist on a spectrum, from evasive violations like hate speech to overt ones like misinformation.
- Misinformation is tricky because it's not intentionally evasive, yet it spreads widely and causes significant harm.