Apache Falcon is a data processing and management solution for Hadoop designed for data motion, coordination of data pipelines, lifecycle management, and data discovery. Falcon enables end consumers to quickly onboard their data and its associated processing and management tasks on Hadoop clusters.
... [More]
Data Management on Hadoop encompasses data motion, process orchestration, lifecycle management, data discovery, etc. among other concerns.
Falcon will enable easy data management via declarative mechanism for Hadoop. Users of Falcon platform simply define infrastructure endpoints, data sets and processing rules declaratively. This information about inter-dependencies between various entities allows Falcon to orchestrate and manage various data management functions. [Less]
Concourse is a pipeline-based CI system written in Go. Rather than a myriad of checkboxes, pipelines are defined as a single declarative config file.
Tagline: CI that scales with your project.
libpipeline is a C library for setting up and running pipelines of processes, without needing to involve shell command-line parsing which is often error-prone and insecure.
Vent is a self-contained operating system based on boot2docker that provides a platform to collect and analyze data across a flexible set of tools and technologies.
This site uses cookies to give you the best possible experience.
By using the site, you consent to our use of cookies.
For more information, please see our
Privacy Policy