Use Airflow to author workflows as directed acyclic graphs (DAGs) of tasks. The Airflow scheduler executes your tasks on an array of workers while following the specified dependencies. Rich command lines utilities makes performing complex surgeries on DAGs a snap. The rich user interface makes it easy to visualize pipelines running in production, monitor progress and troubleshoot issues when needed. | It aims to provide a unified interface for constructing and managing workflows on different workflow engines, such as Argo Workflows, Tekton Pipelines, and Apache Airflow. |
Dynamic: Airflow pipelines are configuration as code (Python), allowing for dynamic pipeline generation. This allows for writting code that instantiate pipelines dynamically.;Extensible: Easily define your own operators, executors and extend the library so that it fits the level of abstraction that suits your environment.;Elegant: Airflow pipelines are lean and explicit. Parameterizing your scripts is built in the core of Airflow using powerful Jinja templating engine.;Scalable: Airflow has a modular architecture and uses a message queue to talk to orchestrate an arbitrary number of workers. Airflow is ready to scale to infinity. | Unified interface and imperative programming style for defining workflows with automatic construction of directed acyclic graph (DAG);
Extensible to support various workflow engines;
Reusable steps for tasks such as distributed training of machine learning models;
Automatic workflow and resource optimizations under the hood |
Statistics | |
GitHub Stars - | GitHub Stars 943 |
GitHub Forks - | GitHub Forks 88 |
Stacks 1.7K | Stacks 0 |
Followers 2.8K | Followers 2 |
Votes 128 | Votes 0 |
Pros & Cons | |
Pros
Cons
| No community feedback yet |
Integrations | |
| No integrations available | |

It makes it easy to automate all your software workflows, now with world-class CI/CD. Build, test, and deploy your code right from GitHub. Make code reviews, branch management, and issue triaging work the way you want.

It implements batch and streaming data processing jobs that run on any execution engine. It executes pipelines on multiple execution environments.

Developer framework to orchestrate multiple services and APIs into your software application using logic triggered by events and time. Build ETL processes, A/B testing, real-time alerts and personalized user experiences with custom logic.

It is a Python module that helps you build complex pipelines of batch jobs. It handles dependency resolution, workflow management, visualization etc. It also comes with Hadoop support built in.

Build and map powerful workflows across tools to save your team time. No coding required. Create rules to define what information flows between each of your tools, in minutes.

na

Camunda enables organizations to operationalize and automate AI, integrating human tasks, existing and future systems without compromising security, governance, or innovation.

It is an organizational tool that makes life easier. It's a surprisingly powerful way to take notes, make lists, collaborate, brainstorm, plan and generally organize your brain.

It is a microservice orchestration platform which enables developers to build scalable applications without sacrificing productivity or reliability. Temporal server executes units of application logic, workflows, in a resilient manner that automatically handles intermittent failures, and retries failed operations.

It is a server-based workflow scheduling system to manage Hadoop jobs. Workflows in it are defined as a collection of control flow and action nodes in a directed acyclic graph. Control flow nodes define the beginning and the end of a workflow as well as a mechanism to control the workflow execution path.