Deploy Workloads with Databricks Workflows
Contact us to book this courseData Engineering
On-Site, Virtual
1 day
By scheduling tasks with Databricks Jobs, applications can be run automatically to keep tables in the Lakehouse fresh. Using Databricks SQL to schedule updates to queries and dashboards allows quick insights using the newest data. In this course, students will be introduced to task orchestration using the Databricks Workflow Jobs UI. Optionally, they will configure and schedule dashboards and alerts to reflect updates to production data pipelines.
In this half-day course, you’ll learn how to orchestrate data pipelines with Databricks Workflow Jobs and schedule dashboard updates to keep analytics up-to-date. We’ll cover topics like getting started with Databricks Workflows, how to use Databricks SQL for on-demand queries, and how to configure and schedule dashboards and alerts to reflect updates to production data pipelines.
Objectives
- By the end of this course, attendees should be able to:
-
Orchestrate tasks with Databricks Workflow Jobs.
-
Use Databricks SQL for on-demand queries.
-
Configure and schedule dashboards and alerts to reflect updates to production data pipelines.
-
Prerequisites
-
Beginner-level familiarity with basic cloud concepts (virtual machines, object storage, identity management)
-
Ability to perform basic code development tasks (create compute, run code in notebooks, use basic notebook operations, import repos from git, etc)
-
Intermediate familiarity with basic SQL concepts (CREATE, SELECT, INSERT, UPDATE, DELETE, WHILE, GROUP BY, JOIN, etc.)
Course outline
- Scheduling Tasks with the Jobs UI
- Workflows Lab
- Explore Scheduling Options
- Conditional Tasks and Repairing Runs
- Modular Orchestration