Skip to main content

Get Started with Databricks for Machine Learning

In this course, you will develop the foundational skills needed to use the Databricks Data Intelligence Platform for executing basic machine learning workflows and supporting data science workloads. You will explore the platform from the perspective of a machine learning practitioner, covering topics such as feature engineering with Databricks Notebooks and model lifecycle tracking with MLflow. Additionally, you will learn about real-time model inference with Mosaic AI Model Serving and experience Databricks’ “glass box” approach to model development through AutoML. The course includes three instructor-led demonstrations, culminating in a comprehensive lab that reinforces the concepts covered in the demos.


Languages Available: English | 日本語 | Português BR | 한국어

Skill Level
Onboarding
Duration
2h
Prerequisites
  • A beginner-level understanding of Python.

  • Basic understanding of DS/ML concepts (e.g. classification and regression models), common model metrics (e.g. F1-score), and Python libraries (e.g. scikit-learn and XGBoost). 

Outline

Databricks Overview

  • Databricks Data Intelligence Platform
  • Demo: Databricks Workspace Walkthrough


Using Databricks for Machine Learning

  • Introduction to Machine Learning with Databricks
  • Exploratory Data Analysis (EDA) and Feature Engineering on Databricks
  • Demo: EDA and Feature Engineering
  • Introduction to MLflow on Databricks
  • Demo: Tracking and Managing Models with MLflow
  • Introduction to Mosaic AI AutoML
  • Demo: Experimentation with Mosaic AI AutoML
  • Introduction to Mosaic AI Model Serving
  • DemoGetting Started with Mosaic AI Model Serving
  • Comprehensive Lab: Getting Started with Databricks for ML

Upcoming Public Classes

Date
Time
Language
Price
Nov 03
09 AM - 11 AM (America/Los_Angeles)
English
Free
Nov 07
12 PM - 02 PM (Asia/Singapore)
English
Free
Dec 12
09 AM - 11 AM (America/Los_Angeles)
English
Free
Dec 15
03 PM - 05 PM (Europe/London)
English
Free
Jan 05
09 AM - 11 AM (America/Los_Angeles)
English
Free
Jan 14
12 PM - 02 PM (Asia/Singapore)
English
Free

Public Class Registration

If your company has purchased success credits or has a learning subscription, please fill out the Training Request form. Otherwise, you can register below.

Private Class Request

If your company is interested in private training, please submit a request.

See all our registration options

Registration options

Databricks has a delivery method for wherever you are on your learning journey

Runtime

Self-Paced

Custom-fit learning paths for data, analytics, and AI roles and career paths through on-demand videos

Register now

Instructors

Instructor-Led

Public and private courses taught by expert instructors across half-day to two-day courses

Register now

Learning

Blended Learning

Self-paced and weekly instructor-led sessions for every style of learner to optimize course completion and knowledge retention. Go to Subscriptions Catalog tab to purchase

Purchase now

Scale

Skills@Scale

Comprehensive training offering for large scale customers that includes learning elements for every style of learning. Inquire with your account executive for details

Upcoming Public Classes

Data Engineer

Build Data Pipelines with Lakeflow Declarative Pipelines

This course introduces users to the essential concepts and skills needed to build data pipelines using Lakeflow Declarative Pipelines in Databricks for incremental batch or streaming ingestion and processing through multiple streaming tables and materialized views. Designed for data engineers new to Lakeflow Declarative Pipelines, the course provides a comprehensive overview of core components such as incremental data processing, streaming tables, materialized views, and temporary views, highlighting their specific purposes and differences.

Topics covered include:

- Developing and debugging ETL pipelines with the multi-file editor in Lakeflow using SQL (with Python code examples provided)

- How Lakeflow Declarative Pipelines track data dependencies in a pipeline through the pipeline graph

- Configuring pipeline compute resources, data assets, trigger modes, and other advanced options

Next, the course introduces data quality expectations in Lakeflow, guiding users through the process of integrating expectations into pipelines to validate and enforce data integrity. Learners will then explore how to put a pipeline into production, including scheduling options, and enabling pipeline event logging to monitor pipeline performance and health.

Finally, the course covers how to implement Change Data Capture (CDC) using the AUTO CDC INTO syntax within Lakeflow Declarative Pipelines to manage slowly changing dimensions (SCD Type 1 and Type 2), preparing users to integrate CDC into their own pipelines.

Note: Databricks Academy is transitioning from video lectures to a more streamlined PDF format with slides and notes for all self-paced courses. Please note that demo videos will still be available in their original format. We would love to hear your thoughts on this change, so please share your feedback through the course survey at the end. Thank you for being a part of our learning community!

Languages Available: English | 日本語 | Português BR | 한국어

Free
2h
Associate

Questions?

If you have any questions, please refer to our Frequently Asked Questions page.