# Apache airflow quick start link: https://airflow.apache.org/docs/stable/start.html # export the AIRFLOW_HOME vi ~/.bash_profile # setting AIRFLOW HOME export AIRFLOW_HOME=/User/Desktop/airflow/ cd ~AIRFLOW_HOME # start the virtual environment python3 -m venv ./venv # to show the list of dependencies pip3 list # install apache airflow pip3 install apache-airflow # initialize the airflow database $ airflow initdb # starting the webserver on port 8080 $ airflow webserver -p 8080 Now, we must be able to see Airflow-DAG's on local URL : http://localhost:8080/admin/ # start the scheduler $ airflow scheduler # Try to review the airflow config file found under AIRFLOW_HOME dir or go to UI and then follow the Admin -> Configuration menu. $ cat airflow.cfg We can learn more about airflow features from the configuration files as below: It can store logs remotely in AWS S3 , Google Cloud Storage or Elastic Search ( remote_logs , j...
ETL, big-data, cloud(AWS/GCP/Azure) technologies and possibly share random stuff along the way!!