Hadoop Cluster

What Does Hadoop Cluster Mean?

A Hadoop cluster is a hardware cluster used to facilitate utilization of open-source Hadoop technology for data handling. The cluster consists of a group of nodes, which are processes running on either a physical or virtual machine. The Hadoop cluster works in coordination to deal with unstructured data and produce data results.


Techopedia Explains Hadoop Cluster

The Hadoop cluster works on a master/slave model. A node called NameNode is the Hadoop master. This node communicates with various DataNode nodes in the cluster to support operations. Hadoop clusters typically also use other Apache open-source technologies like Apache MapReduce and Apache Yarn – the Yarn scheduler helps to direct collaborative activity by various nodes in the system, which may be running on virtual machines or containers. In general, Hadoop clusters are used for all sorts of enterprise technologies and predictive analytics, product and service development, customer relationship management and much more.


Related Terms

Margaret Rouse

Margaret is an award-winning technical writer and teacher known for her ability to explain complex technical subjects to a non-technical business audience. Over the past twenty years, her IT definitions have been published by Que in an encyclopedia of technology terms and cited in articles by the New York Times, Time Magazine, USA Today, ZDNet, PC Magazine, and Discovery Magazine. She joined Techopedia in 2011. Margaret's idea of a fun day is helping IT and business professionals learn to speak each other’s highly specialized languages.