Apache DolphinScheduler is the modern data orchestration platform. Agile to create high performance workflow with low-code
-
Updated
May 9, 2023 - Java
Apache DolphinScheduler is the modern data orchestration platform. Agile to create high performance workflow with low-code
An orchestration platform for the development, production, and observation of data assets.
Build data pipelines, the easy way
MLeap: Deploy ML Pipelines to Production
Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.
Open-source data observability for analytics engineers.
Extract & Load with joy — CLI & version control for ELT without limitations. No more black box. Let your creativity flow.
First open-source data discovery and observability platform. We make a life for data practitioners easy so you can focus on your business.
Optimus is an easy-to-use, reliable, and performant workflow orchestrator for data transformation, data modeling, pipelines, and data quality management.
Dataform is a framework for managing SQL based data operations in BigQuery, Snowflake, and Redshift
The best place to learn data engineering. Built and maintained by the data engineering community.
Build, run and manage your data pipelines with Python or SQL on any cloud
Recap tracks and transform schemas across your whole application.
Data anomalies monitoring as dbt tests and dbt artifacts uploader.
Developer platform for production ML.
Pipebird is open source infrastructure for securely sharing data with customers.
Relational data pipelines for the science lab
Dataplane is an Airflow inspired data platform with additional data mesh capability to automate, schedule and design data pipelines and workflows. Dataplane is written in Golang with a React front end.
A lightweight CLI tool for versioning data alongside source code and building data pipelines.
Add a description, image, and links to the data-pipelines topic page so that developers can more easily learn about it.
To associate your repository with the data-pipelines topic, visit your repo's landing page and select "manage topics."