In this course, the students will implement various data platform technologies into solutions that are in line with business and technical requirements including on-premises, cloud, and hybrid data scenarios incorporating both relational and No-SQL data. They will also learn how to process data using a range of technologies and languages for both streaming and batch data. Students will also explore how to implement data security including authentication, authorization, data policies and standards. They will also define and implement data solution monitoring for both the data storage and data processing activities. Finally, they will manage and troubleshoot Azure data solutions which includes the optimization and disaster recovery of big data, batch processing and streaming data solutions.
Pass the DP-200 and DP-201 exams to be awarded the Microsoft Certified: Azure Data Engineer Associate certification.
Students learn how to develop data solutions using Azure technologies.
- Implement Azure cloud data warehouses
- Implement No-SQL Databases
- Implement Azure SQL Database
- Implement hybrid data scenarios
- Manage Azure DevOps Pipelines
- Implement big data environments
- Develop batch processing solutions
- Develop streaming solutions
- Develop integration solutions
- Implement data migration
- Automate Data Factory Pipelines
- Manage source data access security
- Configure authentication and authorization
- Manage and enforce data policies and standards
- Set up notifications
- Monitor data storage
- Monitor databases for a specified scenario
- Monitor data processing
- Manage Optimization (HIVE, Data Lake, SPARK)
- Manage business continuity
- Azure fundamentals
Module 1: Implement data storage solutions
This module teaches the variety of ways to store data in Azure. The Student will learn the basics of storage management in Azure, how to create a Storage Account, and how to choose the right model for the data you want to store in the cloud. They will also understand how data lake storage can be created to support a wide variety of big data analytics solutions with minimal effort. Learn how to work with NoSQL data using Azure Cosmos DB. They will learn how to provision the service, and how they can load and interrogate data in the service using Visual Studio Code extensions, and the Azure Cosmos DB .NET Core SDK. They will also learn how to configure the availability options so that users are able to access the data from anywhere in the world.
Module 2: Manage and develop data processing
In this module, students will explore the Azure relational data platform options including SQL Database and SQL Data Warehouse. The student will be able explain why they would choose one service over another, and how to provision, connect and manage each of the services. Also, learn how Azure Data factory can be used to orchestrate the data movement and transformation from a wide range of data platform technologies. They will be able to explain the capabilities of the technology and be able to set up an end to end data pipeline that ingests and transforms data.
Module 3: Manage data security
In this module, students will learn how Azure Storage provides a multi-layered security model to protect your data. The students will explore how security can range from setting up secure networks and access keys, to defining permission through to monitoring with Advanced Threat Detection.
Module 4: Monitor data solutions
In this module, students will learn the concepts of event processing and streaming data and how this applies to Events Hubs and Azure Stream Analytics. The students will then set up a stream analytics job to stream data and learn how to query the incoming data to perform analysis of the data. Finally, you will learn how to manage and monitor running jobs.
Module 5: Manage and troubleshoot Azure data solutions
In this module, the student will look at the wide range of monitoring capabilities that are available to provide operational support should there be issue with a data platform architecture. They will explore the data engineering troubleshooting approach and be able to apply this to common data storage and data processing issues.
Dedicated instructor-led training is designed for group training and is delivered by the experts at Opsgility. Delivery availability is anywhere in the world at your location or using advanced virtual training software.
- Standard or Customized Curriculum
- Globally Available for Delivery
- Holistic Learning Plans are Available
- Industry Recognized Subject Matter Experts