Missouri University of Science and Technology, Rolla, USA.
International Journal of Science and Research Archive, 2025, 17(03), 1037-1043
Article DOI: 10.30574/ijsra.2025.17.3.3114
Received on 08 October 2025; revised on 18 November 2025; accepted on 20 November 2025
The growing intricacy and scope of data processing systems have raised the significance of multi-purpose and clever ETL pipelines (Extract, Transform, Load) to the state of deliberations. The data-typical integration has been switched to real-time data integration, which at times makes the self-healing of ETL workflow a requirement. The paper includes the description of the design philosophy, architecture, and the process of practice of self-healing ETL pipelines creation with the help of Apache Airflow and Databricks. It provides a clue of how the ETL systems are going to transform themselves in the recent past to be event-driven and AI-enhanced pipes in the cloud and serverless worlds. It is concerned with alerts in a fault, automated recovery, generative AI-assisted, and distributed architecture-assisted pipeline adaptivity. The review also includes modern techniques and emerging technologies, and this has helped ETL systems to automatically detect, troubleshoot, and remediate failure and the resultant effect is low downtime and the result load.
Self-Healing ETL; Airflow; Databricks; Pipeline Automation
Get Your e Certificate of Publication using below link
Preview Article PDF
Jayanth Veeramachaneni. Designing self-healing ETL pipelines with airflow and databricks. International Journal of Science and Research Archive, 2025, 17(03), 1037-1043. Article DOI: https://doi.org/10.30574/ijsra.2025.17.3.3114.
Copyright © 2025 Author(s) retain the copyright of this article. This article is published under the terms of the Creative Commons Attribution Liscense 4.0







