Apache Airflow is an open source tool that can be used to programmatically author, schedule and monitor data pipelines using Python and SQL. Created at Airbnb as an open-source project in 2014, Airflow was brought into the Apache Software Foundation’s Incubator Program 2016 and announced as Top-Level Apache Project in 2019. It is used as a data orchestration solution, with over 140 integrations and community support.
N/A
Ansible
Score 9.2 out of 10
N/A
The Red Hat Ansible Automation Platform (acquired by Red Hat in 2015) is a foundation for building and operating automation across an organization. The platform includes tools needed to implement enterprise-wide automation, and can automate resource provisioning, and IT environments and configuration of systems and devices. It can be used in a CI/CD process to provision the target environment and to then deploy the application on it.
$5,000
per year
Pricing
Apache Airflow
Red Hat Ansible Automation Platform
Editions & Modules
No answers on this topic
Basic Tower
5,000
per year
Enterprise Tower
10,000
per year
Premium Tower
14,000
per year
Offerings
Pricing Offerings
Apache Airflow
Ansible
Free Trial
No
No
Free/Freemium Version
Yes
No
Premium Consulting/Integration Services
No
No
Entry-level Setup Fee
No setup fee
No setup fee
Additional Details
—
—
More Pricing Information
Community Pulse
Apache Airflow
Red Hat Ansible Automation Platform
Features
Apache Airflow
Red Hat Ansible Automation Platform
Workload Automation
Comparison of Workload Automation features of Product A and Product B
Apache Airflow
9.8
10 Ratings
17% above category average
Red Hat Ansible Automation Platform
-
Ratings
Multi-platform scheduling
10.010 Ratings
00 Ratings
Central monitoring
10.010 Ratings
00 Ratings
Logging
10.010 Ratings
00 Ratings
Alerts and notifications
10.010 Ratings
00 Ratings
Analysis and visualization
10.010 Ratings
00 Ratings
Application integration
9.010 Ratings
00 Ratings
Configuration Management
Comparison of Configuration Management features of Product A and Product B
Airflow is well-suited for data engineering pipelines, creating scheduled workflows, and working with various data sources. You can implement almost any kind of DAG for any use case using the different operators or enforce your operator using the Python operator with ease. The MLOps feature of Airflow can be enhanced to match MLFlow-like features, making Airflow the go-to solution for all workloads, from data science to data engineering.
For automating the configuration of a multi-node, multi-domain (Storage, VM, Container) cluster, Ansible is still the best choice; however, it is not an easy task to achieve. Creating the infrastructure layer, i.e., creating network nodes, VMs, and K8s clusters, still can't be achieved via Ansible. Additionally, error handling remains complex to resolve.
Apache Airflow is one of the best Orchestration platforms and a go-to scheduler for teams building a data platform or pipelines.
Apache Airflow supports multiple operators, such as the Databricks, Spark, and Python operators. All of these provide us with functionality to implement any business logic.
Apache Airflow is highly scalable, and we can run a large number of DAGs with ease. It provided HA and replication for workers. Maintaining airflow deployments is very easy, even for smaller teams, and we also get lots of metrics for observability.
Debugging is easy, as it tells you exactly within your job where the job failed, even when jumping around several playbooks.
Ansible seems to integrate with everything, and the community is big enough that if you are unsure how to approach converting a process into a playbook, you can usually find something similar to what you are trying to do.
Security in AAP seems to be pretty straightforward. Easy to organize and identify who has what permissions or can only see the content based on the organization they belong to.
UI/Dashboard can be updated to be customisable, and jobs summary in groups of errors/failures/success, instead of each job, so that a summary of errors can be used as a starting point for reviewing them.
Navigation - It's a bit dated. Could do with more modern web navigation UX. i.e. sidebars navigation instead of browser back/forward.
Again core functional reorg in terms of UX. Navigation can be improved for core functions as well, instead of discovery.
I can't think of any right now because I've heard about the Lightspeed and I'm really excited about that. Ansible has been really solid for us. We haven't had any issues. Maybe the upgrade process, but other than that, as coming from a user, it's awesome.
Even is if it's a great tool, we are looking to renew our licence for our production servers only. The product is very expensive to use, so we might look for a cheaper solution for our non-production servers. One of the solution we are looking, is AWX, free, and similar to AAP. This is be perfect for our non-production servers.
For its capability to connect with multicloud environments. Access Control management is something that we don't get in all the schedulers and orchestrators. But although it provides so many flexibility and options to due to python , some level of knowledge of python is needed to be able to build workflows.
It's overall pretty easy to use foe all the applications I've mentioned before: configuring hosts, installing packages through tools like apt, applying yaml, making changes across wide groups of hosts, etc. Its not a 10 because of the inconveinience of the yaml setup, and the time to write is not worth it for something applied one time to only a few hosts
Great in almost every way compared to any other configuration management software. The only thing I wish for is python3 support. Other than that, YAML is much improved compared to the Ruby of Chef. The agentless nature is incredibly convenient for managing systems quickly, and if a member of your term has no terminal experience whatsoever they can still use the UI.
There is a lot of good documentation that Ansible and Red Hat provide which should help get someone started with making Ansible useful. But once you get to more complicated scenarios, you will benefit from learning from others. I have not used Red Hat support for work with Ansible, but many of the online resources are helpful.
Multiple DAGs can be orchestrated simultaneously at varying times, and runs can be reproduced or replicated with relative ease. Overall, utilizing Apache Airflow is easier to use than other solutions now on the market. It is simple to integrate in Apache Airflow, and the workflow can be monitored and scheduling can be done quickly using Apache Airflow. We advocate using this tool for automating the data pipeline or process.
AAP compares favorably with Terraform and Power Automate. I don't have much experience with Terraform, but I find AAP and Ansible easier to use as well as having more capabilities. Power Platform is also an excellent automation tool that is user friendly but I feel that Ansible has more compatibility with a variety of technologies.
Impact Depends on number of workflows. If there are lot of workflows then it has a better usecase as the implementation is justified as it needs resources , dedicated VMs, Database that has a cost
POSITIVE: currently used by the IT department and some others, but we want others to use it.
NEGATIVE: We need less technical output for the non-technical. It should be controllable or a setting within playbooks. We also need more graphical responses (non-technical).
POSITIVE: Always being updated and expanded (CaC, EDA, Policy as Code, execution environments, AI, etc..)