Published: July 14, 2025
by Ying Chen, Megha Agarwal, Asfandyar Qureshi and Qi Zheng
We’re excited to announce that Google's Gemma 3 models are coming to Databricks, starting with the Gemma 3 12B model, which is now natively available across all clouds. Text capabilities are available today, with multimodal support rolling out soon.
Gemma 3 12B is optimized for enterprise workloads, striking the ideal balance between capability and computational efficiency. It excels at core use cases like document processing, content analysis, code generation, and conversational AI, making it a strong fit for production-grade applications.
Databricks has long been the platform where enterprises manage and analyze unstructured data at scale. As enterprises connect that data with large language models to build AI agents, the need for efficient, high-quality models with a reasonable price point has grown rapidly. Gemma 3 12B fills a critical gap—offering open, high-quality multimodal capabilities that power document AI and visual question answering use cases. Combined with Databricks’ unified platform for unstructured data and model development, teams can build and deploy production-grade AI faster and more affordably.
Gemma 3 12B provides an attractive balance of size and quality:
To get a sense of whether Gemma 3 12B would suit your use case, try it out in AI Playground.
You can query the model serving endpoint as well. For both AI Playground and the model serving endpoint, multimodal capabilities are coming soon.
Additionally, the newly released MLflow 3 allows you to evaluate the model more comprehensively across your specific datasets.
You can also run scalable batch inference by sending a SQL query to your table.