Hadoop


The Apache™ Hadoop™ project develops open-source software for reliable, scalable, distributed computing. The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using a simple programming model. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-avaiability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-availabile service on top of a cluster of computers, each of which may be prone to failures.

Courtesy of Apache Hadoop

Following mailing lists have been archived for Hadoop:

hadoop-common-dev
hadoop-common-user
hadoop-common-issues
hadoop-common-commits
hadoop-mapreduce-dev
hadoop-hdfs-dev
hadoop-mapreduce-user
hadoop-general
hadoop-mapreduce-commits
hadoop-hdfs-commits
hadoop-hdfs-user
scm-users
cdh-user
Welcome to QnaList Q&A, where you can search answers from other members of the community.