Airflow Github

Apache Airflow tutorial. Contribute to tuanavu/airflow-tutorial development by creating an account on GitHub.

  1. Docker Compose Logic Broken. Init Not Happening - Github.com
  2. Airflow Github
  3. Cached

Docker Compose Logic Broken. Init Not Happening - Github.com

Airflow Github

Airflow Github

  • Installation tools ¶. The official way of installing Airflow is with the pip tool. There was a recent (November 2020) change in resolver, so currently only 20.2.4 version is officially supported, although you might have a success with 20.3.3+ version (to be confirmed if all initial issues from pip 20.3.0 release have been fixed in 20.3.3).
  • Apache Airflow is an open-source workflow management platform. It started at Airbnb in October 2014 as a solution to manage the company's increasingly complex workflows. Creating Airflow allowed Airbnb to programmatically author and schedule their workflows and monitor them via the built-in Airflow user interface. From the beginning, the project was made open source, becoming an Apache Incubator.
airflow-quick-start.sh
# install
mkdir ~/airflow
cd~/airflow
pip install airflow
# Have a look here if you need additional packages: https://airflow.incubator.apache.org/installation.html
# setup mysql backend as given here. The default SQLite is not adequate for some workloads.
# http://site.clairvoyantsoft.com/installing-and-configuring-apache-airflow/
# start services
mysql.server start
nohup airflow webserver $*>>~/airflow/logs/webserver.logs &
nohup airflow scheduler >>~/airflow/logs/scheduler.logs &
# setup mysql
CREATE USER 'airflow'@'localhost' IDENTIFIED BY 'airflow';
GRANT ALL PRIVILEGES ON *.* TO 'airflow'@'localhost' WITH GRANT OPTION;
airflow initdb
# write your awesome dags. Airflow tutorial is just awesome!
# https://airflow.incubator.apache.org/tutorial.html
# to see current dags:
airflow list_dags
# to test your dag:
# sanity check:
python ~/airflow/dags/dag_contents_dir/dag_your_aweseome_workflow.py
# list and test each task inside your dag named 'your-awesome-workflow':
airflow list_tasks your-awesome-workflow
airflow test your-awesome-workflow your-awesome-workflow-task YYYY-MM-DD
# manual trigger
airflow trigger_dag -e YYYY-MM-DD your-awesome-workflow
# to manually trigger externally triggered dag
airflow trigger_dag --conf {'file_variable': '/path/to/file'} dag_id
airflow trigger_dag -c '{'batch': 'YYYYMMDD'}' your-awesome-externally-triggered-workflow
# to update packages
pip install --upgrade google-api-python-client
# to reset db
airflow resetdb
# shutdown services:
kill$(ps -ef grep 'airflow webserver' awk '{print $2}')
kill$(ps -ef grep 'airflow scheduler' awk '{print $2}')
# in order to re-run a dag run:
# 1. From airflow UI, delete all the task instances for the dag run
# 2. From airflow UI, change state of the dag run instance to ** running **

Cached

Sign up for freeto join this conversation on GitHub. Already have an account? Sign in to comment