
Data Engineering Podcast Warehouse Native Incremental Data Processing With Dynamic Tables And Delayed View Semantics
20 snips
Jul 21, 2025 Dan Sotolongo, a principal engineer at Snowflake, shares insights on simplifying data engineering through incremental data processing and delayed view semantics. He dives into the complexities of managing evolving datasets in cloud warehouses, discussing how these concepts optimize resource use and reduce latency. The conversation contrasts traditional batch systems with dynamic tables and streaming solutions, emphasizing the need for a unified framework for semantic guarantees in data pipelines, and highlights the ongoing innovations in data integration and maintenance.
AI Snips
Chapters
Transcript
Episode notes
Dynamic Tables Enhance Workflow Simplicity
- Dynamic tables ease building and running incremental pipelines but don't change the underlying schemas or data models.
- They enable running pipelines continuously with resource isolation, improving operational simplicity.
Handling Input and Immutability
- Iceberg and open table formats help unify data input/output, enabling Snowflake to sidestep integration complexity.
- Immutability features extend dynamic tables beyond view semantics to handle deletions and GDPR constraints.
Continuous Data Validation Approach
- Monitor data with datametric functions and alert on anomalies to ensure data quality in continuous pipelines.
- Build logic to isolate bad rows in dynamic tables for later inspection and correction.
