Explore how to work with the TaskFlow API, perform operations using TaskFlow, integrate PostgreSQL in Airflow, use sensors in Airflow, and work with hooks in Airflow.
Overview
Syllabus
Introduction
- Apache Airflow essential training
- Prerequisites
- Quick Airflow setup overview
- DAG using PythonOperators
- DAG using TaskFlow
- Passing data using XCom with operators
- Passing data using the TaskFlow API
- Tasks with multiple outputs
- Passing multiple outputs in TaskFlow
- Branching using operators
- Branching using the TaskFlow API
- Complete branching pipeline using TaskFlow
- Interoperating and passing data between operators and TaskFlow
- Performing SQL operations with TaskFlow
- Data transformation and storage using TaskFlow
- Installing PostgreSQL on macOS
- Installing PostgreSQL on WSL
- Connecting to PostgreSQL
- Using the PostgreSQL operator
- Performing PostgreSQL insert operations
- Performing PostgreSQL join operations
- A complete end-to-end pipeline with PostgreSQL
- Configuring PostgreSQL as a metadata database and using the LocalExecutor
- Introducing the FileSensor
- Creating a pipeline using the FileSensor
- Running DAGs with a FileSensor
- Creating and running a pipeline with SQLSensors
- Setting up for a PostgreSQL pipeline with hooks
- Creating and running a pipeline with PostgresSQL hooks
- Setting up access to Amazon S3 buckets
- Setting up a connection to Amazon S3 buckets
- Creating and running a pipeline with an S3 hook
- Summary and next steps
Taught by
Janani Ravi