Event data from IoT, clickstream, and application telemetry powers critical real-time analytics and AI when combined with the Databricks Data Intelligence Platform. Traditionally, ingesting this data required several data hops (message bus, Spark jobs) between the data source and the lakehouse. This adds operational overhead, data duplication, requires specialized expertise, and it's generally inefficient when the lakehouse is the only destination for this data.
Once this data lands in the lakehouse, it is transformed and curated for downstream analytical use cases. However, teams need to serve this analytical data for operational use cases, and building these custom applications can be a laborious process. They need to provision and maintain essential infrastructure components like a dedicated OLTP database instance (with networking, monitoring, backups, and more). Additionally, they need to manage the reverse ETL process for the analytical data into the database to resurface it in a real-time application. Customers also often build additional pipelines to push data from the lakehouse into these external operational databases. These pipelines add to the infrastructure that developers need to set up and maintain, altogether diverting their attention from the main goal: building the applications for their business.
So how does Databricks simplify both ingesting data into the lakehouse and serving gold data to support operational workloads?
Enter Zerobus Ingest and Lakebase.
Zerobus Ingest, part of Lakeflow Connect, is a set of APIs that provide a streamlined way to push event data directly into the lakehouse. Eliminating the single-sink message bus layer entirely, Zerobus Ingest reduces infrastructure, simplifies operations, and delivers near real-time ingestion at scale. As such, Zerobus Ingest makes it easier than ever to unlock the value of your data.
The data-producing application must specify a target table to write data to, ensure that the messages map correctly to the table's schema, and then initiate a stream to send data to Databricks. On the Databricks side, the API validates the schemas of the message and the table, writes the data to the target table, and sends an acknowledgment to the client that the data has been persisted.
Zerobus Ingest Capability | Specifications |
Ingestion latency | Near real-time (≤5 seconds) |
Max throughput per client | Up to 100 MB/sec |
Concurrent clients | Thousands per table |
Continuous sync lag (Delta → Lakebase) | 10–15 seconds |
Real-time foreach writer latency | 200–300 milliseconds |
Lakebase is a fully managed, serverless, scalable, Postgres database built into the Databricks Platform, designed for low-latency operational and transactional workloads that run directly on the same data powering analytical and AI use cases.
The complete separation of compute and storage delivers rapid provisioning and elastic autoscaling. Lakebase's integration with the Databricks Platform is a major differentiator from traditional databases because Lakebase makes Lakehouse data directly available to both real-time applications and AI without the need for complex custom data pipelines. It is built to deliver database creation, query latency, and concurrency requirements to power enterprise applications and agentic workloads. Lastly, it allows developers to easily version control and branch databases like code.
Together, these tools allow customers to ingest data from several systems directly into Delta tables and implement reverse ETL use cases at scale. Next, we will explore how to use these technologies to implement a near real-time application!
As a practical example, let's help ‘Data Diners,' a food delivery company, empower their management staff with an application to monitor driver activity and order deliveries in real-time. Currently, they lack this visibility, which limits their ability to mitigate issues as they arise during deliveries.
Why is a real-time application valuable?
Let's see how to build this with Zerobus Ingest, Lakebase, and Databricks Apps on the Data Intelligence Platform!

This end-to-end architecture follows four stages: (1) A data producer uses the Zerobus SDK to write events directly to a Delta table in Databricks Unity Catalog. (2) A continuous sync pipeline pushes updated records from the Delta table to a Lakebase Postgres instance. (3) A FastAPI backend connects to Lakebase via WebSockets to stream real-time updates. (4) A front-end application built on Databricks Apps visualizes the live data for end users.
Starting with our data producer, the data diner app on the driver's phone will emit GPS telemetry data about the driver's location (latitude and longitude coordinates) en route to deliver orders. This data will be sent to an API gateway, which ultimately sends the data to the next service in the ingestion architecture.
With the Zerobus SDK, we can quickly write a client to forward events from the API gateway to our target table. With the target table being updated in near real time, we can then create a continuous sync pipeline to update our lakebase tables. Finally, by leveraging Databricks Apps, we can deploy a FastAPI backend that uses WebSockets to stream real-time updates from Postgres, along with a front-end application to visualize the live data flow.
Before the introduction of the Zerobus SDK, the streaming architecture would have included multiple hops before it landed in the target table. Our API gateway would have needed to offload the data to a staging area like Kafka, and we would need Spark Structured Streaming to write the transactions into the target table. All of this adds unnecessary complexity, especially given that the sole destination is the lakehouse. The architecture above instead demonstrates how the Databricks Data Intelligence Platform simplifies end-to-end enterprise application development — from data ingestion to real-time analytics and implementation of interactive applications.
Prerequisites: What You Need
The event data produced by the client applications will live in a Delta table. Use the code below to create that target table in your desired catalog and schema.
The code below pushes the telemetry events data into Databricks using the Zerobus API.
As of today, Zerobus Ingest does not support CDF. CDF allows Databricks to record change events for new data written to a delta table. These change events could be inserts, deletes, or updates. These change events can then be used to update the synced tables in Lakebase. To sync data to Lakebase and continue with our project, we will write the data in the target table to a new table and enable CDF on that table.
To power the app, we will sync data from this new, CDF-enabled table into a Lakebase instance. We will sync this table continuously to support our near real-time dashboard.

In the UI, we select:
This ensures the app reflects the latest data with minimal delay.
Note: You can also create the sync pipeline programmatically using the Databricks SDK.
Continuous syncs from Delta to Lakebase has a 10-15-second lag, so if you need lower latency, consider using real-time mode via ForeachWriter writer to sync data directly from a DataFrame to a Lakebase table. This will sync the data within milliseconds.
Refer to the Lakebase ForeachWriter code on Github.

With your data synced to Lakebase, you can now deploy your code to build your app. In this example, the app fetches events data from Lakebase and uses it to update a near real-time application to track a driver’s activity while en route to making food deliveries. Read the Get Started with Databricks Apps docs to learn more about building apps on Databricks.
Check out more tutorials, demos and solution accelerators to build your own applications for your specific needs.
Learn more about Zerobus Ingest, Lakebase, and Databricks Apps in the technical documentation. You can also take a look at the Databricks Apps Cookbook and Cookbook Resource Collection.
IoT, clickstream, telemetry, and similar applications generate billions of data points every day, which are used to power critical real-time applications across several industries. As such, simplifying ingestion from these systems is paramount. Zerobus Ingest provides a streamlined way to push event data directly from these systems into the lakehouse while ensuring high performance. It pairs nicely with Lakebase to simplify end-to-end enterprise application development.
