BigData / Apache Airflow Interview Questions
What are best practices for writing efficient Airflow DAGs?
Key best practices for production-quality DAGs:
- Keep DAG files lightweight — avoid heavy imports or database calls at parse time; the scheduler parses DAG files continuously.
- Use top-level constants only — don't call APIs or read files at module level; do it inside operators/callables.
- Set catchup=False unless backfilling is intentional.
- Prefer TaskFlow API for clarity and automatic XCom passing.
- Use sensors in reschedule mode for long waits.
- Keep tasks idempotent and atomic.
- Use Pools to protect downstream systems from overload.
- Set email_on_failure and SLAs for alerting.
- Avoid using Variables at top-level — each call hits the DB at parse time.
Invest now in Acorns!!! 🚀
Join Acorns and get your $5 bonus!
Acorns is a micro-investing app that automatically invests your "spare change" from daily purchases into diversified, expert-built portfolios of ETFs. It is designed for beginners, allowing you to start investing with as little as $5. The service automates saving and investing. Disclosure: I may receive a referral bonus.
Invest now!!! Get Free equity stock (US, UK only)!
Use Robinhood app to invest in stocks. It is safe and secure. Use the Referral link to claim your free stock when you sign up!.
The Robinhood app makes it easy to trade stocks, crypto and more.
Webull! Receive free stock by signing up using the link: Webull signup.
More Related questions...
