Skip to main content
<
Page 2
>
Industries category icon 1

US Air Force Hackathon: How Large Language Models Will Revolutionize USAF Flight Test

[DISTRIBUTION STATEMENT A. Approved for public release; Distribution is unlimited 412TW-PA-24004] The views expressed are those of the author and do not reflect...
Generative AI

OLMo Is Here, Powered by Databricks

February 1, 2024 by Jonathan Frankle in Mosaic Research
As Chief Scientist (Neural Networks) at Databricks, I lead our research team toward the goal of giving everyone the ability to build and...
Generative AI

Serving Quantized LLMs on NVIDIA H100 Tensor Core GPUs

Quantization is a technique for making machine learning models smaller and faster. We quantize Llama2-70B-Chat, producing an equivalent-quality model that generates 2.2x more...
Generative AI

Building and Customizing GenAI with Databricks: LLMs and Beyond

Generative AI has opened new worlds of possibilities for businesses and is being emphatically embraced across organizations. According to a recent MIT Tech...
Generative AI

LLM Training and Inference with Intel Gaudi 2 AI Accelerators

January 4, 2024 by Abhi Venigalla and Daya Khudia in Mosaic Research
At Databricks, we want to help our customers build and deploy generative AI applications on their own data without sacrificing data privacy or...
Generative AI

Integrating NVIDIA TensorRT-LLM with the Databricks Inference Stack

Over the past six months, we've been working with NVIDIA to get the most out of their new TensorRT-LLM library. TensorRT-LLM provides an easy-to-use Python interface to integrate with a web server for fast, efficient inference performance with LLMs. In this post, we're highlighting some key areas where our collaboration with NVIDIA has been particularly important.
Generative AI

Patronus AI: Using LLMs to Detect Business-Sensitive Information

November 1, 2023 by Emily Hutson in Mosaic Research
EnterprisePII is a first-of-its-kind large language model (LLM) data set aimed at detecting business-sensitive information. The challenge of detecting and redacting sensitive business...
Generative AI

Training LLMs at Scale with AMD MI250 GPUs

October 30, 2023 by Abhi Venigalla in Mosaic Research
Introduction Four months ago, we shared how AMD had emerged as a capable platform for generative AI and demonstrated how to easily and...
Generative AI

LLM Training on Unity Catalog data with MosaicML Streaming Dataset

Introduction Large Language Models (LLMs) have given us a way to generate text, extract information, and identify patterns in industries from healthcare to...
Generative AI

LLM Inference Performance Engineering: Best Practices

In this blog post, the MosaicML engineering team shares best practices for how to capitalize on popular open source large language models (LLMs)...