Introduction to Lakeflow Spark Declarative Pipelines: Reliable Data Pipelines Made Easy
Overview
| Experience | In Person |
|---|---|
| Track | Data Engineering & Streaming |
| Industry | Healthcare & Life Sciences, Manufacturing, Financial Services |
| Technologies | Lakeflow |
| Skill Level | Beginner |
Building data pipelines often means hand-writing complex data engineering patterns like incremental processing, CDC/SCD handling, and backfills in every job.Lakeflow Spark Declarative Pipelines (SDP) are the next evolution in how transformations are written in Spark by default. You declare the data you want to produce, and the platform applies proven execution semantics for you. Built on an open, declarative standard for Spark pipelines, SDP helps teams move faster on hard problems while scaling production best practices consistently.In this session, you’ll see how SDP:
- Provides a necessary foundation for agentic data pipeline development
- Reduces the amount of custom logic required for incremental and CDC workloads
- Makes batch and streaming pipelines easier to build, debug, and evolve over time
- Standardizes reliability and correctness across pipelines as they scale
We’ll walk through concrete examples that show how declarative pipelines simplify otherwise complex data engineering patterns.
Session Speakers
Zoé Durand
/Staff Product Manager
Databricks
Ray Zhu
/Director, Product Management
Databricks