Since 2022, Databricks SQL (DBSQL) Serverless has delivered a 5x performance gain across real-world customer workloads—turning a 100-second dashboard into a 20-second one. That acceleration came from continuous engine improvements, all delivered automatically and without performance tuning.
Today, we’re adding even more. With the launch of Predictive Query Execution and Photon Vectorized Shuffle, queries get up to 25% faster on top of the existing 5x gains, bringing that 20-second dashboard down to around 15 seconds. These new engine improvements roll out automatically across all DBSQL Serverless warehouses, at zero additional cost
When it launched in Apache Spark, Adaptive Query Execution (AQE) was a big step forward. It allowed queries to re-plan based on actual data sizes as the query was executed. However, it had one major limitation: it could only act after a query execution stage was completed. That delay meant problems like data skew or excessive spilling often weren’t caught until it was too late.
Predictive Query Execution (PQE) changes that. It introduces a continuous feedback loop inside the query engine:
The result? Faster queries, fewer surprises, and more predictable performance—especially for complex pipelines and mixed workloads
Photon is a native C++ engine that processes data in columnar batches, vectorized to leverage modern CPUs and execute SQL queries several times faster. Shuffle operations, which restructure large datasets between stages, remain among the heaviest in query processing.
Shuffle operations historically are the hardest type to optimize because they involve lots of random memory access. It’s also rarely possible to reduce the number of random accesses without rewriting the data. The key intuition that we had was that instead of reducing the number of random accesses, we could reduce the distance between each random access in memory.
This led to us rewriting Photon's shuffle from the ground up with column-based Shuffle for higher cache and memory efficiency.
The result is a shuffle component that moves data efficiently, executes fewer instructions, and considers cache. With the newly optimized shuffle, we see 1.5× higher throughput in CPU-bound workloads like large joins.
This upgrade is rolling out now across all DBSQL Serverless warehouses—no action needed.
Haven’t tried DBSQL Serverless yet? Now’s the perfect time. Serverless is the easiest way to run analytics on the Lakehouse:
Just create a DBSQL Serverless warehouse and start querying—zero tuning required. If you are not already using Databricks SQL, read more on enabling serverless SQL warehouses.