Richard Zhang, senior research scientist at Adobe Research, discusses perceptual metrics and the LPIPS paper, detection tools for fake visual content, and data attribution and concept ablation in generative AI. They explore challenges in visual generative AI, improving perceptual metrics and loss functions, controllability of generative AI systems, addressing challenges in the ecosystem, understanding the connection between synthesized images and training data, and concept ablation and opting out in the contributor domain.
Read more
AI Summary
Highlights
AI Chapters
Episode notes
auto_awesome
Podcast summary created with Snipd AI
Quick takeaways
Richard Zhang's work on perceptual metrics like LPIPS improves human-computer alignment in generative AI.
Addressing the challenges of detecting fake visual content is crucial for maintaining trust in generative AI.
Deep dives
Training image-based generative AI models
Richard Zhang discusses his work on image-based generative AI, specifically focusing on the use of deep networks for image generation. He highlights the challenges faced in predicting high-dimensional signals and the limitations of existing loss functions. Zhang shares his early experiences in generative tasks, such as image colorization, and the shift towards adding controllability to generative AI systems.
Improving perceptual metrics and loss functions
Zhang explores the importance of creating better loss functions that align with human perception. He discusses the limitations of using the L2 or Euclidean distance as a loss function and presents the concept of learned perceptual image patch similarity metric (LPIPS). Zhang explains how data-driven approaches, such as collecting human judgments on distorted image pairs, can be used to evaluate and improve the perceptual quality of generative models.
Visual generative AI as an ecosystem
Zhang describes visual generative AI as an ecosystem involving creators, consumers, and contributors. He highlights the need to address the visual quality bottleneck in generative AI and emphasizes the importance of providing tools for consumers to discern between real and synthetic content. Additionally, Zhang discusses the significance of acknowledging and compensating data contributors, and the potential for an open standard for recording metadata to track the origin and subsequent edits of visual assets.
Detection and attribution in generative AI
Zhang explores the challenges and approaches in detecting manipulated or synthetic images in the context of generative AI. He discusses the use of data-driven techniques for forensic analysis, including the simulation of future methods to test robustness and the effectiveness of data augmentation. Zhang also highlights the importance of a multi-pronged approach, combining provenance, detection, and public education to address the challenges posed by malicious actors.
Today we’re joined by Richard Zhang, senior research scientist at Adobe Research. In our conversation with Richard, we explore the research challenges that arise when regarding visual generative AI from an ecosystem perspective, considering the disparate needs of creators, consumers, and contributors. We start with his work on perceptual metrics and the LPIPS paper, which allow us to better align human perception and computer vision and which remain used in contemporary generative AI applications such as stable diffusion, GANs, and latent diffusion. We look at his work creating detection tools for fake visual content, highlighting the importance of generalization of these detection methods to new, unseen models. Lastly, we dig into his work on data attribution and concept ablation, which aim to address the challenging open problem of allowing artists and others to manage their contributions to generative AI training data sets.
The complete show notes for this episode can be found at twimlai.com/go/656.
Get the Snipd podcast app
Unlock the knowledge in podcasts with the podcast player of the future.
AI-powered podcast player
Listen to all your favourite podcasts with AI-powered features
Discover highlights
Listen to the best highlights from the podcasts you love and dive into the full episode
Save any moment
Hear something you like? Tap your headphones to save it with AI-generated key takeaways
Share & Export
Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more
AI-powered podcast player
Listen to all your favourite podcasts with AI-powered features
Discover highlights
Listen to the best highlights from the podcasts you love and dive into the full episode