

Warehouse Native Incremental Data Processing With Dynamic Tables And Delayed View Semantics
Jul 21, 2025
Dan Sotolongo, a principal engineer at Snowflake, shares insights on simplifying data engineering through incremental data processing and delayed view semantics. He dives into the complexities of managing evolving datasets in cloud warehouses, discussing how these concepts optimize resource use and reduce latency. The conversation contrasts traditional batch systems with dynamic tables and streaming solutions, emphasizing the need for a unified framework for semantic guarantees in data pipelines, and highlights the ongoing innovations in data integration and maintenance.
AI Snips
Chapters
Transcript
Episode notes
Dan's Data Engineering Origin Story
- Dan Sotolongo began coding in middle school and studied physics and computer science in undergrad.
- Early experience with data analysis for physics experiments shaped his interest in data engineering.
Essence of Incremental Processing
- Incremental data processing minimizes work by updating only changed parts of continuously evolving datasets.
- This reduces resource costs and latency compared to reprocessing entire datasets every time.
Dynamic Tables Blend Streams and Batch
- Batch systems are easier to maintain but less fresh; streaming systems offer low latency but complex operations.
- Dynamic tables offer a hybrid with streaming programming model and micro-batch engine for easier use and typical latency needs.