The Apache? Hadoop? project develops open-source software for reliable, scalable, distributed computing.
The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster of computers, each of which may be prone to failures.
Apache Hadoop 项目旨在开发高可靠、可扩展的分布式计算开源软件。
Apache Hadoop软件库是一个框架,它使用简单的编程模型分布式处理跨计算机集群大型数据集。该框架能够从单个服务器扩展到数千台机器,并且每台机器提供本地计算和存储。与其依赖硬件来提供高可用性,库本身的设计目的是在应用层检测和处理故障,因此可以在计算机集群上提供高可用服务,同时集群中的每台计算机都可能发生故障。