Starting today, you can use Google’s Gemini models natively and securely on the Databricks Data Intelligence Platform. This marks a major milestone for enterprise AI: Databricks now offers secure, unified access to all of the world’s top LLMs, right where your data lives.
As part of our Week of Agents, this release expands how customers can build, govern, and deploy powerful AI agents securely at scale, bringing the latest Gemini models into the same trusted, governed environment as your data and workflows.
With this release, any team can now:
This makes it even easier to deploy advanced conversational agents, automate document analysis, and accelerate business reasoning by applying the latest Gemini models safely and efficiently right where your data is stored in Databricks.
Gemini models generally available in Databricks Foundation Model API, which allows users to call Gemini models hosted on Vertex AI, from within SQL, as an API endpoint in Model Serving, or via Agent Bricks.
One way to leverage Gemini models on Databricks is as a built-in operator in SQL or Python. This dramatically simplifies the process for applying LLMs directly to enterprise data and can automate routine tasks like analyzing contracts, PDFs, transcripts or images. When you run these queries, Databricks automatically scales Gemini models capacity in the backend to handle everything from a handful of rows to millions, ensuring fast and reliable results without extra setup.
Figure 1: Try Gemini models with ai_query in your workspace today!
Additionally, Gemini models are available via our real-time APIs at scale! You can use either the OpenAI chat completions client or our REST API.
Figure 2: Use Gemini models to build a real-time agent with tools in Python
You can now access two of Google's most advanced models, Gemini-2.5 Pro and Gemini 2.5 Flash, directly from the Data Intelligence Platform in Databricks.
Gemini 2.5 Flash is distinct in the market for blending the highest levels of intelligence with incredibly low latency and high throughput. This model is one of Google’s hybrid reasoning model, designed to “think before it speaks” and allows developers to set the level of “thinking” according to the task. It excels at structured problem solving, tool use (like Python or calculators), and step-by-step logic.

Use Gemini 2.5 Flash when you need to:
Gemini 2.5 Pro is one of Google’s most capable large language models. It provides state of the art performance in advanced reasoning, industry leading 1M context window for long-context processing, coding, and is natively multi-modal.

Use Gemini 2.5 Pro when you need to:
Real-time customer support agents
With Gemini 2.5 Flash, enterprises can build chatbots that respond in milliseconds while pulling in enterprise data securely. For example, a telco company can deliver automated support that classifies an issue, retrieves account information, and suggests a fix—all before a human agent needs to step in.
Multi-modal product intelligence
Gemini 2.5 Pro enables workflows that combine images, text, and structured data. A retailer can analyze product photos, user reviews, and inventory data together to detect defects or predict sales trends.
Decision automation at enterprise scale
Using Databricks orchestration and governance, organizations can build agents that run thousands of structured reasoning tasks per minute—such as categorizing transactions, scoring risks, or generating compliance reports—balancing Flash for latency with Pro for accuracy.
