Tags
Language
Tags
December 2024
Su Mo Tu We Th Fr Sa
1 2 3 4 5 6 7
8 9 10 11 12 13 14
15 16 17 18 19 20 21
22 23 24 25 26 27 28
29 30 31 1 2 3 4

The Complete Hands-On Introduction To Airbyte

Posted By: ELK1nG
The Complete Hands-On Introduction To Airbyte

The Complete Hands-On Introduction To Airbyte
Published 3/2024
MP4 | Video: h264, 1920x1080 | Audio: AAC, 44.1 KHz
Language: English | Size: 1.79 GB | Duration: 3h 16m

Get started with Airbyte and learn how to use it with Apache Airflow, Snowflake, dbt and more

What you'll learn

Understand what Airbyte is, its architecture, concepts, and its role in the MDS

Install and set up Airbyte locally with Docker

Connect Airbyte to different data sources (databases, cloud storages, etc)

Configure Airbyte to send data to various destinations (DWs, databases)

Develop a data pipeline from scratch with Airbyte, dbt, Soda, Airflow, Postgres, and Snowflake to run your first data syncs

Set up monitoring and notifications with Airbyte

Requirements

Prior experience with Python

Access to Docker on a local machine

A Google Cloud account with a billing account (for BigQuery)

Description

Welcome to the Complete Hands-On Introduction to Airbyte!Airbyte is an open-source data integration engine that helps you consolidate data in your data warehouses, lakes, and databases. It is an alternative to Stich and Fivetran and provides hundreds of connectors mainly built by the community. Aibyte has many connectors (+300) and is extensible. You can create your connector if it doesn't exist. In this course, you will learn everything you need to get started with Airbyte:What is Airbyte? Where does it fit in the data stack, and why it is helpful for you.Essential concepts such as source, destination, connections, normalization, etc.How to create a source and a destination to synchronize data at ease.Airbyte best practices to efficiently move data between endpoints.How to set up and run Airbyte locally with Docker and KubernetesBuild a data pipeline from scratch using Airflow, dbt, Postgres, Snowflake, Airbyte and Soda.And more.At the end of the course, you will fully understand Airbyte and be ready to use it with your data stack!If you need any help, don't hesitate to ask in Q/A section of Udemy, I will be more than happy to help!See you in the course!

Overview

Section 1: Welcome!

Lecture 1 Welcome!

Lecture 2 Prerequisites

Lecture 3 Who am I?

Lecture 4 Learning Advice

Section 2: Airbyte Fundamentals

Lecture 5 Why Airbyte?

Lecture 6 What is Airbyte?

Lecture 7 The Core Concepts

Lecture 8 The Core Components

Lecture 9 Why not Airbyte?

Lecture 10 Airbyte Cloud or OSS?

Section 3: Getting started with Airbyte

Lecture 11 Introduction to Docker (Optional)

Lecture 12 Running Airbyte with Docker

Lecture 13 The Airbyte UI tour

Lecture 14 The Bank Pipeline

Lecture 15 Create your first source (Google Sheets)

Lecture 16 Create your first destination (BigQuery)

Lecture 17 Configure your first connection

Lecture 18 Make your first sync!

Lecture 19 Raw tables and additional columns?

Lecture 20 Connector classifications (Certified, Community, etc)

Section 4: Advanced Concepts

Lecture 21 How does a sync work?

Lecture 22 Side notes for Postgres

Lecture 23 Discover the sync modes

Lecture 24 Handling schema changes

Lecture 25 What is Change Data Capture (CDC)?

Lecture 26 Enable CDC with Postgres

Lecture 27 Syncing data between Postgres and BigQuery using CDC

Lecture 28 The Sync Modes cheat sheet

Lecture 29 CDC under the hood

Section 5: The Fraud Project

Lecture 30 Project overview

Lecture 31 Learning recommendations

Lecture 32 The Setup and requirements

Lecture 33 The Data Generators (Python Scripts)

Lecture 34 Quick introduction to Apache Airflow

Lecture 35 Let's generate some data!

Lecture 36 Set up the S3 bucket with the user

Lecture 37 Create the AWS S3 destination with Airbyte

Lecture 38 Create the Postgres to S3 connection with Airbyte

Lecture 39 Create the MySQL source with Airbyte

Lecture 40 Create the MySQL to S3 connection with Airbyte

Lecture 41 What's the Write Audi Publish pattern?

Lecture 42 Create the AWS S3 source with Airbyte

Lecture 43 Create the Snowflake destination with Airbyte

Lecture 44 Create the Raw to Staging connection with Airbyte

Lecture 45 Quick introduction to Soda

Lecture 46 Write data quality checks with Soda

Lecture 47 Create the customer_metrics table with dbt

Lecture 48 Create the Fraud Data Pipeline with Airflow

Lecture 49 The pipeline in action!

Data Engineers,Analytics Engineers,Data Architects