Apache Spark Fundamentals
This training course delivers key concepts and methods for data processing applications development using Apache Spark.
To be determined
The Azure Data Services course is designed to offer participants a comprehensive understanding of the various data services available within Microsoft Azure, both relational (e.g., Azure SQL Database, PostgreSQL) and non-relational (e.g., Blob Storage, Cosmos DB, Event Hub). Modules cover storage architecture, management, replication, and specific use cases for each service type. Additionally, the course includes practical instruction on using Azure Data Factory to automate and streamline data transfer between different services, enabling participants to design robust data integration pipelines. This training is ideal for backend developers, DBAs, data engineers, and architects aiming to leverage Azure’s diverse data services for efficient data management.
The course begins with an introduction to Azure Data Services and their architectures, followed by an overview of certification paths, pricing, and monitoring. Participants then dive into the specifics of relational storages like Azure SQL Database and PostgreSQL, where they explore data operations, partitioning, replication, and best practices for various use cases.
For non-relational storage, modules cover Blob Storage, Azure Data Lake Storage (ADLS) Gen 1/2, and Cosmos DB, along with caching solutions such as Redis. Each module introduces core functionalities, including read/write operations, replication, and integration scenarios. Participants also learn about Azure Event Hub for real-time data streaming and utilize Azure Data Factory to automate data flows between relational and non-relational storages. Practical exercises reinforce these concepts, helping participants master the diverse capabilities of Azure Data Services.
Upon completion, participants will be able to:
The course is structured with 60% theory and 40% hands-on practice. Labs focus on configuring storage solutions, performing data operations, and designing integration workflows with Azure Data Factory, ensuring practical application of each Azure data service.
- Azure Data Services Certification
- Azure Data Services Overview
- Azure Data Services Architecture
- Azure Data Services Pricing
- Azure Data Services Monitoring
- Azure Data Services Integration
- Starting with Azure Database for PostgreSQL
- Accessing Azure Database for PostgreSQL
- DDL (Data Definition Language) and DML (Data Manipulation Language) with SQL
- Read path
- Write path
- Partitioning
- Replication
- Use cases
- Starting with Azure SQL Database
- Accessing Azure SQL Database
- DDL (Data Definition Language) and DML (Data Manipulation Language) with SQL
- Read path
- Write path
- Partitioning
- Replication
- Use cases
- Starting with Azure Blob Storage
- Accessing Azure Blob Storage
- Upload/Download files
- Read path
- Write path
- Replication
- Use cases
- Why we need Azure ADLS Gen 1/2
- Starting with Azure ADLS Gen 1/2
- Accessing Azure ADLS Gen 1/2
- Upload/Download files
- Read path
- Write path
- Partitioning
- Replication
- Starting with Azure Data Factory
- Accessing Azure Data Factory
- Creating a pipeline
- Creating a dataset
- Creating a linked service
- Creating a trigger
- Starting with Azure Event Hub
- Accessing Azure Event Hub
- Creating a topic
- Writing to a topic
- Reading from a topic
- Replication
- Use cases
- Starting with Azure Data Factory
- Accessing Azure Data Factory
- Creating a pipeline
- Starting with Azure Cosmos DB
- Accessing Azure Cosmos DB
- Creating a database
- Creating a table
- Writing to a table
- Reading from a table
- Replication
- Use cases
- Starting and accessing with Azure Redis Cache
- Creating, Writing and Reading a cache
- Replication
- Use cases
- Starting and accessing with Azure Table Storage
- Creating, Writing and Reading a table
- Replication
- Use cases
- Starting and accessing with Azure File Storage
- Creating, Writing and Reading a file
- Replication
- Use cases