
Data Engineering on Microsoft Azure (DP-203T00)
In this course, the student will learn how to implement and manage data engineering workloads on Microsoft Azure, using Azure services such as Azure Synapse Analytics, Azure Data Lake Storage Gen2, Azure Stream Analytics, Azure Databricks, and others. The course focuses on common data engineering tasks such as orchestrating data transfer and transformation pipelines, working with data files in a data lake, creating and loading relational data warehouses, capturing and aggregating streams of real-time data, and tracking data assets and lineage.
The primary audience for this course is data professionals, data architects, and business intelligence professionals who want to learn about data engineering and building analytical solutions using data platform technologies that exist on Microsoft Azure. The secondary audience for this course includes data analysts and data scientists who work with analytical solutions built on Microsoft Azure.
- Explore core data engineering concepts and the capabilities of Azure Data Lake Storage Gen2 for big data workloads.
- Query, transform, and manage data using Azure Synapse serverless SQL pools, lake databases, and pipelines.
- Analyze and transform data with Apache Spark and Delta Lake in Azure Synapse and Azure Databricks.
- Design and load data into relational data warehouses and optimize their performance.
- Implement security, authentication, and permissions in Azure Synapse Analytics.
- Ingest and process real-time streaming data with Azure Stream Analytics and Power BI.
- Plan and support hybrid transactional and analytical processing (HTAP) using Azure Synapse Link.
Webucator is a Microsoft Certified Partner. This class uses official Microsoft courseware and will be delivered by a Microsoft Certified Trainer (MCT).
- Introduction to data engineering on Azure
- What is data engineering
- Important data engineering concepts
- Data engineering in Microsoft Azure
- Introduction to Azure Data Lake Storage Gen2
- Understand Azure Data Lake Storage Gen2
- Enable Azure Data Lake Storage Gen2 in Azure Storage
- Compare Azure Data Lake Store to Azure Blob storage
- Understand the stages for processing big data
- Use Azure Data Lake Storage Gen2 in data analytics workloads
- Introduction to Azure Synapse Analytics
- What is Azure Synapse Analytics
- How Azure Synapse Analytics works
- When to use Azure Synapse Analytics
- Use Azure Synapse serverless SQL pool to query files in a data lake
- Understand Azure Synapse serverless SQL pool capabilities and use cases
- Query files using a serverless SQL pool
- Create external database objects
- Use Azure Synapse serverless SQL pools to transform data in a data lake
- Transform data files with the CREATE EXTERNAL TABLE AS SELECT statement
- Encapsulate data transformations in a stored procedure
- Include a data transformation stored procedure in a pipeline
- Create a lake database in Azure Synapse Analytics
- Understand lake database concepts
- Explore database templates
- Create a lake database
- Use a lake database
- Secure data and manage users in Azure Synapse serverless SQL pools
- Choose an authentication method in Azure Synapse serverless SQL pools
- Manage users in Azure Synapse serverless SQL pools
- Manage user permissions in Azure Synapse serverless SQL pools
- Analyze data with Apache Spark in Azure Synapse Analytics
- Get to know Apache Spark
- Use Spark in Azure Synapse Analytics
- Analyze data with Spark
- Visualize data with Spark
- Transform data with Spark in Azure Synapse Analytics
- Modify and save dataframes
- Partition data files
- Transform data with SQL
- Use Delta Lake in Azure Synapse Analytics
- Understand Delta Lake
- Create Delta Lake tables
- Create catalog tables
- Use Delta Lake with streaming data
- Use Delta Lake in a SQL pool
- Build a data pipeline in Azure Synapse Analytics
- Understand pipelines in Azure Synapse Analytics
- Create a pipeline in Azure Synapse Studio
- Define data flows
- Run a pipeline
- Use Spark Notebooks in an Azure Synapse Pipeline
- Understand Synapse Notebooks and Pipelines
- Use a Synapse notebook activity in a pipeline
- Use parameters in a notebook
- Introduction to Azure Synapse Analytics
- What is Azure Synapse Analytics
- How Azure Synapse Analytics works
- When to use Azure Synapse Analytics
- Use Azure Synapse serverless SQL pool to query files in a data lake
- Understand Azure Synapse serverless SQL pool capabilities and use cases
- Query files using a serverless SQL pool
- Create external database objects
- Analyze data with Apache Spark in Azure Synapse Analytics
- Get to know Apache Spark
- Use Spark in Azure Synapse Analytics
- Analyze data with Spark
- Visualize data with Spark
- Use Delta Lake in Azure Synapse Analytics
- Understand Delta Lake
- Create Delta Lake tables
- Create catalog tables
- Use Delta Lake with streaming data
- Use Delta Lake in a SQL pool
- Analyze data in a relational data warehouse
- Design a data warehouse schema
- Create data warehouse tables
- Load data warehouse tables
- Query a data warehouse
- Build a data pipeline in Azure Synapse Analytics
- Understand pipelines in Azure Synapse Analytics
- Create a pipeline in Azure Synapse Studio
- Define data flows
- Run a pipeline
- Analyze data in a relational data warehouse
- Design a data warehouse schema
- Create data warehouse tables
- Load data warehouse tables
- Query a data warehouse
- Load data into a relational data warehouse
- Load staging tables
- Load dimension tables
- Load time dimension tables
- Load slowly changing dimensions
- Load fact tables
- Perform post load optimization
- Manage and monitor data warehouse activities in Azure Synapse Analytics
- Scale compute resources in Azure Synapse Analytics
- Pause compute in Azure Synapse Analytics
- Manage workloads in Azure Synapse Analytics
- Use Azure Advisor to review recommendations
- Use dynamic management views to identify and troubleshoot query performance
- Secure a data warehouse in Azure Synapse Analytics
- Understand network security options for Azure Synapse Analytics
- Configure Conditional Access
- Configure authentication
- Manage authorization through column and row level security
- Manage sensitive data with Dynamic Data Masking
- Implement encryption in Azure Synapse Analytics
- Plan hybrid transactional and analytical processing using Azure Synapse Analytics
- Understand hybrid transactional and analytical processing patterns
- Describe Azure Synapse Link
- Implement Azure Synapse Link with Azure Cosmos DB
- Enable Cosmos DB account to use Azure Synapse Link
- Create an analytical store enabled container
- Create a linked service for Cosmos DB
- Query Cosmos DB data with Spark
- Query Cosmos DB with Synapse SQL
- Implement Azure Synapse Link for SQL
- What is Azure Synapse Link for SQL?
- Configure Azure Synapse Link for Azure SQL Database
- Configure Azure Synapse Link for SQL Server 2022
- Get started with Azure Stream Analytics
- Understand data streams
- Understand event processing
- Understand window functions
- Ingest streaming data using Azure Stream Analytics and Azure Synapse Analytics
- Stream ingestion scenarios
- Configure inputs and outputs
- Define a query to select, filter, and aggregate data
- Run a job to ingest data
- Visualize real-time data with Azure Stream Analytics and Power BI
- Use a Power BI output in Azure Stream Analytics
- Create a query for real-time visualization
- Create real-time data visualizations in Power BI
- Explore Azure Databricks
- Get started with Azure Databricks
- Identify Azure Databricks workloads
- Understand key concepts
- Data governance using Unity Catalog and Microsoft Purview
- Perform data analysis with Azure Databricks
- Ingest data with Azure Databricks
- Data exploration tools in Azure Databricks
- Data analysis using DataFrame APIs
- Use Apache Spark in Azure Databricks
- Get to know Spark
- Create a Spark cluster
- Use Spark in notebooks
- Use Spark to work with data files
- Visualize data
- Manage data with Delta Lake
- Get started with Delta Lake
- Create Delta tables
- Implement schema enforcement
- Data versioning and time travel in Delta Lake
- Data integrity with Delta Lake
- Build Lakeflow Declarative Pipelines
- Explore Lakeflow Declarative Pipelines
- Data ingestion and integration
- Real-time processing
- Deploy workloads with Lakeflow Jobs
- What are Lakeflow Jobs?
- Understand key components of Lakeflow Jobs
- Explore the benefits of Lakeflow Jobs
- Deploy workloads using Lakeflow Jobs
Each student will receive a comprehensive set of materials, including course notes and all the class examples.
Experience in the following is required for this Azure class:
- Knowledge of cloud computing and core data concepts and professional experience with data solutions.
- Comfort with basic Python and SQL.
- Familiarity with various Azure services including Azure Synapse Analytics, Azure Stream Analytics, Azure Cosmos DB, and Azure Databricks.
Courses that can help you meet these prerequisites:
Live Private Class
- Private Class for your Team
- Live training
- Online or On-location
- Customizable
- Expert Instructors