Multiple BashOperators
Airflow DAGs can contain many operators, each performing their defined tasks.
You've successfully implemented one of your scripts as an Airflow task and have decided to continue migrating your individual scripts to a full Airflow DAG. You now want to add more components to the workflow. In addition to the cleanup.sh
used in the previous exercise you have two more scripts, consolidate_data.sh
and push_data.sh
. These further process your data and copy to its final location.
The DAG analytics_dag
is defined (meaning you do not need to add the with DAG(...)
statement, and your cleanup
task is still defined. The BashOperator
is already imported.
This exercise is part of the course
Introduction to Apache Airflow in Python
Exercise instructions
- Define a
BashOperator
calledconsolidate
, to runconsolidate_data.sh
with atask_id
ofconsolidate_task
. - Add a final
BashOperator
calledpush_data
, runningpush_data.sh
and atask_id
ofpushdata_task
.
Hands-on interactive exercise
Have a go at this exercise by completing this sample code.
# Define a second operator to run the `consolidate_data.sh` script
consolidate = ____(
task_id='consolidate_task',
bash_command=____
)
# Define a final operator to execute the `push_data.sh` script
push_data = ____(
____=____,
____=____
)