Session
Data Triggers and Advanced Control Flow With Lakeflow Jobs
Overview
Experience | In Person |
---|---|
Type | Breakout |
Track | Data Engineering and Streaming |
Industry | Enterprise Technology |
Technologies | Databricks Workflows, LakeFlow |
Skill Level | Intermediate |
Duration | 40 min |
Lakeflow Jobs is the production-ready fully managed orchestrator for the entire Lakehouse with 99.95% uptime. Join us for a dive into how you can orchestrate your enterprise data operations, from triggering your jobs only when your data is ready to advanced control flow with conditionals, looping and job modularity — with demos!
Attendees will gain practical insights into optimizing their data operations by orchestrating with Lakeflow Jobs:
- New task types: Publish AI/BI Dashboards, push to Power BI or ingest with Lakeflow Connect
- Advanced execution control: Reference SQL Task outputs, run partial DAGs and perform targeted backfills
- Repair runs: Re-run failed pipelines with surgical precision using task-level repair
- Control flow upgrades: Native for-each loops and conditional logic make DAGs more dynamic + expressive
- Smarter triggers: Kick off jobs based on file arrival or Delta table changes, enabling responsive workflows
- Code-first approach to pipeline orchestration
Session Speakers
IMAGE COMING SOON
Roland Fäustlin
/Product Manager
Databricks