KeerthanaKnowledge Contributor
What are the different types of node in Hadoop?
What are the different types of node in Hadoop?
Sign Up to our social questions and Answers Engine to ask questions, answer people’s questions, and connect with other people.
Login to our social questions & Answers Engine to ask questions answer people’s questions & connect with other people.
Lost your password? Please enter your email address. You will receive a link and will create a new password via email.
Please briefly explain why you feel this question should be reported.
Please briefly explain why you feel this answer should be reported.
Please briefly explain why you feel this user should be reported.
Questions | Answers | Discussions | Knowledge sharing | Communities & more.
NameNode: It’s like the boss that keeps track of where all the files are stored in the Hadoop system.
DataNode: These are the workers that actually store the data on the Hadoop cluster, like shelves in a big storage room.
ResourceManager/NodeManager: They’re like the supervisors that manage resources and tasks in the cluster, making sure everything runs smoothly.
So, you’ve got the boss, the workers, and the supervisors—all working together to handle data storage and processing in Hadoop.
In Hadoop, there are primarily three types of nodes in a cluster:
1. NameNode: This is the master node responsible for managing the file system namespace and metadata. It keeps track of where the data is stored across the cluster and coordinates access to files by clients.
2. DataNode: These are the worker nodes that store the actual data in the Hadoop Distributed File System (HDFS). DataNodes are responsible for storing and retrieving data based on instructions from the NameNode and handling data replication and recovery.
3. ResourceManager/NodeManager: These nodes are part of the Hadoop YARN (Yet Another Resource Negotiator) framework, which manages resources and schedules jobs on the cluster. The ResourceManager is the master node that allocates resources to applications, while the NodeManager is the worker node responsible for managing resources and executing tasks.
These nodes work together to provide distributed storage and processing capabilities in a Hadoop cluster.