Thousands of enterprises already use Llama models on the Databricks Data Intelligence Platform to power AI applications, agents, and workflows. Today, we’re excited to partner with Meta to bring you their latest model series—Llama 4—available today in many Databricks workspaces and rolling out across AWS, Azure, and GCP.
Llama 4 marks a major leap forward in open, multimodal AI—delivering industry-leading performance, higher quality, larger context windows, and improved cost efficiency from the Mixture of Experts (MoE) architecture. All of this is accessible through the same unified REST API, SDK, and SQL interfaces, making it easy to use alongside all your models in a secure, fully governed environment.
The Llama 4 models raise the bar for open foundation models—delivering significantly higher quality and faster inference compared to any previous Llama model.
At launch, we’re introducing Llama 4 Maverick, the largest and highest-quality model from today’s release from Meta. Maverick is purpose-built for developers building sophisticated AI products—combining multilingual fluency, precise image understanding, and safe assistant behavior. It enables:
And you can now build all of this with significantly better performance. Compared to Llama 3.3 (70B), Maverick delivers:
Coming soon to Databricks is Llama 4 Scout—a compact, best-in-class multimodal model that fuses text, image, and video from the start. With up to 10 million tokens of context, Scout is built for advanced long-form reasoning, summarization, and visual understanding.
“With Databricks, we could automate tedious manual tasks by using LLMs to process one million+ files daily for extracting transaction and entity data from property records. We exceeded our accuracy goals by fine-tuning Meta Llama and, using Mosaic AI Model Serving, we scaled this operation massively without the need to manage a large and expensive GPU fleet."