Overview
Implement a Data Analytics Solution with Azure Databricks
Learn how to harness the power of Apache Spark and powerful clusters running on the Azure Databricks platform to run data analytics workloads in a data lakehouse.
Skills Covered
- Explore Azure Databricks
- Use Apache spark in Azure Databricks
- Use Delta Lake in Azure Databricks
- Use SQL Warehouses in Azure Databricks
- Run Azure Databricks Notebooks with Azure Data Factory
Who Should Attend
- IT Professionals
Course Curriculum
Prerequisites
There are no prerequisites required to attend this course.
Download Course Syllabus
Course Modules
Azure Databricks is a cloud service that provides a scalable platform for data analytics using Apache Spark.
Learning objectives
In this module, you’ll learn how to:
- Provision an Azure Databricks workspace.
- Identify core workloads and personas for Azure Databricks.
- Describe key concepts of an Azure Databricks solution.
Prerequisites
Before starting this module, you should have a fundamental knowledge of data analytics concepts. Consider completing Azure Data Fundamentals certification before starting this module.
Azure Databricks is built on Apache Spark and enables data engineers and analysts to run Spark jobs to transform, analyze and visualize data at scale.
Learning objectives
In this module, you’ll learn how to:
- Describe key elements of the Apache Spark architecture.
- Create and configure a Spark cluster.
- Describe use cases for Spark.
- Use Spark to process and analyze data stored in files.
- Use Spark to visualize data.
Prerequisites
Before starting this module, you should have a basic knowledge of Azure Databricks. Consider completing the Explore Azure Databricks module before this one.
Delta Lake is an open source relational storage area for Spark that you can use to implement a data lakehouse architecture in Azure Databricks.
Learning objectives
In this module, you’ll learn how to:
- Describe core features and capabilities of Delta Lake.
- Create and use Delta Lake tables in Azure Databricks.
- Create Spark catalog tables for Delta Lake data.
- Use Delta Lake tables for streaming data.
Prerequisites
Before starting this module, you should know how to use Apache Spark in Azure Databricks. Consider completing the Use Apache Spark in Azure Databricks module before this one.
Azure Databricks provides SQL Warehouses that enable data analysts to work with data using familiar relational SQL queries.
Learning objectives
In this module, you’ll learn how to:
- Create and configure SQL Warehouses in Azure Databricks.
- Create databases and tables.
- Create queries and dashboards.
Prerequisites
Before starting this module, you should have a basic knowledge of Azure Databricks. Consider completing the Explore Azure Databricks module before this one.
Using pipelines in Azure Data Factory to run notebooks in Azure Databricks enables you to automate data engineering processes at cloud scale.
Learning objectives
In this module, you’ll learn how to:
- Describe how Azure Databricks notebooks can be run in a pipeline.
- Create an Azure Data Factory linked service for Azure Databricks.
- Use a Notebook activity in a pipeline.
- Pass parameters to a notebook.
Prerequisites
Before starting this module, you should have a basic knowledge of Azure Databricks. Consider completing the Explore Azure Databricks module before this one.
Request More Information
Training Options
- ILT: Instructor-Led Training
- VILT: Virtual Instructor-Led Training
RM1,200.00Enroll Now
RM1,200.00Enroll Now
RM1,200.00Enroll Now
RM1,200.00Enroll Now
Exam & Certification
This course is not associated with any Certification.