While the buzz around big data analysis is at a peak, there is less discussion about how to get the necessary data into the systems in the first place, which can involve the cumbersome task of setting up and maintaining a number of data processing pipelines.

To help solve this problem, Santa Clara, Calif. start-up DataTorrent has released what it calls the first enterprise-grade ingestion application for Hadoop, DataTorrent dtIngest.

The application is designed to streamline the process of collecting, aggregating and moving data onto and off of a Hadoop cluster.

The software is based on Project Apex, an open source software package available under the Apache 2.0 license.

To read this article in full or to leave a comment, please click here