Implementing a Lakehouse with Microsoft Fabric (DP-601)

Course 8681

  • Duration: 1 day
  • Language: English
  • Level: Intermediate

This course is designed to build your foundational skills in data engineering on Microsoft Fabric, focusing on the Lakehouse concept. This Microsoft Fabric (DP-601) Training course will explore the powerful capabilities of Apache Spark for distributed data processing and the essential techniques for efficient data management, versioning, and reliability by working with Delta Lake tables. This course will also explore data ingestion and orchestration using Dataflows Gen2 and Data Factory pipelines.

This course includes a combination of lectures and hands-on exercises that will prepare you to work with Lakehouses in Microsoft Fabric.

Microsoft Fabric (DP-601) Training Delivery Methods

  • In-Person

  • Online

Microsoft Fabric (DP-601) Training Information

In this course, you will:

  • Discover how Microsoft Fabric can meet your enterprise's analytics needs in one platform.
  • Describe core features and capabilities of lakehouses in Microsoft Fabric.
  • Analyse and process data in a Lakehouse at scale.
  • Create advanced analytics solutions using the enhanced capabilities of delta tables.
  • Visually create multi-step data ingestion and transformation using Power Query Online with Dataflows (Gen2).
  • Create pipelines that orchestrate data ingestion and transformation tasks with Data Factory capabilities within Microsoft Fabric.

Training Prerequisites

You should be familiar with basic data concepts and terminology.

Microsoft Fabric (DP-601) Training Outline

In this module, you'll learn how to:

  • Describe end-to-end analytics in Microsoft Fabric

In this module, you'll learn how to:

  • Describe core features and capabilities of lakehouses in Microsoft Fabric
  • Create a lakehouse
  • Ingest data into files and tables in a lakehouse
  • Query lakehouse tables with SQL

In this module, you'll learn how to:

  • Configure Spark in a Microsoft Fabric workspace
  • Identify suitable scenarios for Spark notebooks and Spark jobs
  • Use Spark dataframes to analyse and transform data
  • Use Spark SQL to query data in tables and views
  • Visualise data in a Spark notebook

In this module, you'll learn how to:

  • Understand Delta Lake and delta tables in Microsoft Fabric
  • Create and manage delta tables using Spark
  • Use Spark to query and transform data in delta tables
  • Use delta tables with Spark structured streaming

In this module, you'll learn how to:

  • Describe Dataflow (Gen2) capabilities in Microsoft Fabric
  • Create Dataflow (Gen2) solutions to ingest and transform data
  • Include a Dataflow (Gen2) in a pipeline

In this module, you'll learn how to:

  • Describe pipeline capabilities in Microsoft Fabric
  • Use the Copy Data activity in a pipeline
  • Create pipelines based on predefined templates
  • Run and monitor pipelines

Need Help Finding The Right Training Solution?

Our training advisors are here for you.

Microsoft Fabric (DP-601) Training FAQs

Microsoft Fabric is an end-to-end analytics solution with full-service capabilities including data movement, data lakes, data engineering, data integration, data science, real-time analytics, and business intelligence—all backed by a shared platform providing robust data security, governance, and compliance.

A lakehouse is a collection of files, folders, and tables that represent a database over a data lake used by the Apache Spark engine and SQL engine for big data processing. A lakehouse includes enhanced capabilities for ACID transactions when using the open-source Delta formatted tables.

The lakehouse item is hosted within a unique workspace folder in Microsoft OneLake. It contains files in various formats (structured and unstructured) organised in folders and subfolders.

The data lake is the foundation on which all the Fabric services are built. Microsoft Fabric Lake is also known as OneLake. It's built into the Fabric service and provides a unified location to store all organisational data where the experiences operate.

  • Data Analysts
  • Data Engineers
  • Data Scientists