Hadoop is a distributed system architecture that is extremely fault-tolerant and designed to be deployed on inexpensive hardware. Function introduction: high reliability. Hadoop's ability to store and process data bit-by-bit is worthy of trust. High scalability. Hadoop distributes data and completes computing tasks among available computer clusters, which can be easily expanded to thousands of nodes. Efficiency. Hadoop can dynamically move data between nodes and ensure the dynamic balance of each node, so the processing speed is very fast. High fault tolerance. Hadoop can automatically save multiple copies of data and automatically redistribute failed tasks. low cost. Compared with all-in-one computers, commercial data warehouses, and data marts such as QlikView and Yonghong Z-Suite, hadoop is open source, so the software cost of the project will be greatly reduced. Hadoop comes with a framework written in the Java language, so it is ideal to run on Linux production platforms. Applications on Hadoop can also be written in other languages, such as C++.
Expand