Introducing Mixtral 8x7B with Databricks Model ServingDecember 21, 2023 by Ahmed Bilal, Daya Khudia, Ankit Mathur, Asfandyar Qureshi, Bruce Fontaine, Linden Li, Maddie Dawson, Sandeep Krishnamurthy, Josh Hartman and Hagay Lupesko in Engineering Blog Today, Databricks is excited to announce support for Mixtral 8x7B in Model Serving . Mixtral 8x7B is a sparse Mixture of Experts (MoE)...
Build GenAI Apps Faster with New Foundation Model CapabilitiesDecember 11, 2023 by Ahmed Bilal, Asfandyar Qureshi, Margaret Qian, Jianwei Xie, Sue Ann Hong, Vladimir Kolovski, Mingyu Li and Ankit Mathur in Engineering Blog Following the announcements we made last week about Retrieval Augmented Generation (RAG) , we're excited to announce major updates to Model Serving. Databricks...
LLM Inference Performance Engineering: Best PracticesOctober 12, 2023 by Megha Agarwal, Asfandyar Qureshi, Nikhil Sardana, Linden Li, Julian Quevedo and Daya Khudia in Mosaic Research In this blog post, the MosaicML engineering team shares best practices for how to capitalize on popular open source large language models (LLMs)...