Hi there-
I'm currently working on getting airflow setup in a distributed and HA setup on AWS and have ran into a few challenges.
Is it possible to tell airflow to use an s3 bucket for its DAG directory and the log directory?
Things I've encountered:
1. Running more than one scheduler introduces race conditions where each scheduler attempts to grab the same task, this is stopped by unique constraints in the DB however it is a nice flurry of errors.
2. If I were to have multiple airflow web servers or schedulers (active passive) or just a new vm, I'd want to preserve my output logs and have them accessible.
Thanks,
Jesse