A distributed system infrastructure developed by the Apache Foundation.
You can develop distributed programs without understanding the details of the distributed underlying layer. Make full use of the power of the cluster for high-speed computing and storage.
[1] hadoop implements a Distributed File System (HDFS. HDFS features high fault tolerance and is designed to be deployed on low-cost hardware. It also provides high throughput to access application data, suitable for applications with large data sets. HDFS relaxed (relax) POSIX requirements and allows you to access data in a streaming Access File System as a stream.
The core design of the hadoop framework is: HDFS and mapreduce. HDFS provide storage for massive data, while mapreduce provides computing for massive data.
The full version of the course class is selfless
:
Http://bbs.timevery.com/bbs/forum.php? MoD = viewthread & tid = 29 & extra =
40 sets of hadoop tutorials for free download and sharing