Lee: AI is just inextricably embedded into social processes, that kind of disciplinary extension still doesn't go far enough. So I'm starting to see a little bit of a turn in some of the criticisms of AI ethics are starting to focus more on these like social issues. And yeah, it seems as emerging fields, AI governance and connectivity, maybe we'll move at the partially towards that away from just purely algorithmic stuff just by necessity.
In episode 57 of The Gradient Podcast, Andrey Kurenkov speaks to Blair Attard-Frost.
Note: this interview was recorded 8 months ago, and some aspects of Canada’s AI strategy have changed since then. It is still a good overview of AI governance and other topics, however.
Blair is a PhD Candidate at the University of Toronto’s Faculty of Information who researches the governance and management of artificial intelligence. More specifically, they are interested in the social construction of intelligence, unintelligence, and artificial intelligence, the relationship between organizational values and AI use, and the political economy, governance, and ethics of AI value chains. They integrate perspectives from service sciences, cognitive sciences, public policy, information management, and queer studies for their research.
Have suggestions for future podcast guests (or other feedback)? Let us know here!
Subscribe to The Gradient Podcast: Apple Podcasts | Spotify | Pocket Casts | RSSFollow The Gradient on Twitter or Mastodon
Outline:
* Intro
* Getting into AI research
* What is AI governance
* Canada’s AI strategy
* Other interests
Links:
* Once a promising leader, Canada’s artificial-intelligence strategy is now a fragmented laggard
* The Ethics of AI Business Practices: A Review of 47 Guidelines
Get full access to The Gradient at
thegradientpub.substack.com/subscribe