At this year’s Data + AI Summit, we introduced the IDE for Data Engineering: a new developer experience purpose-built for authoring data pipelines directly inside the Databricks Workspace. As the new default development experience, the IDE reflects our opinionated approach to data engineering: declarative by default, modular in structure, Git-integrated, and AI-assisted.
In short, the IDE for Data Engineering is everything you need to author and test data pipelines - all in one place.
With this new development experience available in Public Preview, we’d like to use this blog to explain why declarative pipelines benefit from a dedicated IDE experience and highlight the key features that make pipeline development faster, more organized, and easier to debug.
Declarative pipelines simplify data engineering by letting you declare what you want to achieve instead of writing detailed step-by-step instructions on how to build it. Although declarative programming is an extremely powerful approach for building data pipelines, working with multiple datasets and managing the full development lifecycle can become hard to handle without dedicated tooling.
This is why we built a full IDE experience for declarative pipelines directly in the Databricks Workspace. Available as a new editor for Lakeflow Spark Declarative Pipelines, it enables you to declare datasets and quality constraints in files, organize them into folders, and view the connections through an automatically generated dependency graph displayed alongside your code. The editor evaluates your files to determine the most efficient execution plan and allows you to iterate quickly by rerunning single files, a set of changed datasets, or the entire pipeline.
The editor also surfaces execution insights, provides built-in data previews, and includes debugging tools to help you fine-tune your code. It also integrates with version control and scheduled execution with Lakeflow Jobs. Thus, you can perform all tasks related to your pipeline from a single surface.
By consolidating all these capabilities into a single IDE-like surface, the editor enables the practices and productivity data engineers expect from a modern IDE, while staying true to the declarative paradigm.
The video embedded below shows these features in action, with further details covered in the following sections.
"The new editor brings everything into one place - code, pipeline graph, results, configuration, and troubleshooting. No more juggling browser tabs or losing context. Development feels more focused and efficient. I can directly see the impact of each code change. One click takes me to the exact error line, which makes debugging faster. Everything connects - code to data; code to tables; tables to the code. Switching between pipelines is easy, and features like auto-configured utility folders remove complexity. This feels like the way pipeline development should work."— Chris Sharratt, Data Engineer, Rolls-Royce
"In my opinion, the new Pipelines Editor is a huge improvement. I find it much easier to manage complex folder structures and switch between files thanks to the multi-soft tab experience. The integrated DAG view really helps me stay on top of intricate pipelines, and the enhanced error handling is a game changer-it helps me pinpoint issues quickly and streamlines my development workflow."— Matt Adams, Senior Data Platforms Developer, PacificSource Health Plans
We designed the editor so that even users new to the declarative paradigm can quickly build their first pipeline.
These features help users get productive fast, and transition their work into production-ready pipelines.
Building pipelines is an iterative process. The editor streamlines this process with features that simplify authoring and make it faster to test and refine logic:
These capabilities reduce context switching and keep developers focused on building pipeline logic.
Pipeline development involves more than writing code. The new developer experience brings all related tasks onto a single surface, from modularizing code for maintainability to setting up automation and observability:
By unifying these capabilities, the editor streamlines both day-to-day development and long-term pipeline operations.
Check out the video below for more details on all these features in action.
We’re not stopping here. Here’s a preview of what we are currently exploring:
Let us know what else you’d like to see — your feedback drives what we build.
The IDE for data engineering is available in all clouds. To enable it, open a file associated with an existing pipeline, click the ‘Lakeflow Pipelines Editor: OFF’ banner, and toggle it on. You can also enable it during pipeline creation with a similar toggle, or from the User Settings page.
Learn more using these resources:
