Premium

Cloudera Hadoop Developer Certification Questions and Answer (Dumps and Practice Questions)



Question : Which node is considered as slave nodes

  : Which node is considered as slave nodes
1. Secondary NameNode
2. DataNode
3. TaskTracker
4. 1,2 and 3 are correct
5. 2 and 3 are correct




Correct Answer : 5


Explanation: We can consider nodes to be in two different categories:
Master Nodes
- Run the NameNode, Secondary NameNode, JobTracker daemons
- Only one of each of these daemons runs on the cluster

Slave Nodes
- Run the DataNode and TaskTracker daemons
- A slave node will run both of these daemons

Refer HadoopExam.com Recorded Training Module : 2 and 3





Question : Which daemon stores the file data blocks ?

 : Which daemon stores the file data blocks ?
1. NameNode
2. TaskTracker
3. DataNode
4. Secondary Data Node



Correct Answer : 3


Explanation: DataNodes hold the actual blocks
- Each block will be 64MB or 128MB in size
- Each block is replicated three times on the cluster

Refer HadoopExam.com Recorded Training Module : 2 and 16






Question : When a client submits a Job, its configuration information is packaged into XML file

  : When a client submits a Job, its configuration information is packaged into XML file
1. True
2. False



Correct Answer : 1


Explanation: Submitting a Job
- When a client submits a job, its configuration information is packaged into an XML file.

This file, along with the .jar file containing the actual program code, is handed to the JobTracker
- The JobTracker then parcels out individual tasks to TaskTracker nodes
- When a TaskTracker receives a request to run a task, it instantiates a separate JVM for that task
- TaskTracker nodes can be configured to run multiple tasks at the same time
- If the node has enough processing power and memory

Refer HadoopExam.com Recorded Training Module : 3 and 4



Related Questions


Question : Which staement is true about Apache Flume

 :  Which staement is true about Apache Flume
1. Flume is a distributed service
2. it is used to moving large amount of data as it is produced
3. It is ideal for gathering logs from multiple systems
4. Can be used to inserting logs into HDFS
5. All of the above


Question : Which statement is wrong about flume

 :  Which statement is wrong about flume
1. Flume can continue deliver events in the face of system component failure
2. Flume can scales horizontally
3. Flume provides a central Master controller for manageability
4. 1 and 3
5. None of the above



Question :

Flume can be extended by adding connectors to existing storage layers

 :
1. True
2. False




Question :

Select the correct option for the flume architecture

 :
1. The Master holds configuration information for each node
2. Node communicate with the Master in every 5 seconds
3. Nodes passes its version number to master
4. 1 and 3 are correct
5. 1,2 and 3



Question :

Which framework is used for export and import the data between HDFS and RDBMS
 :
1. Flume
2. Sqoop
3. Hive
4. Pig



Question : To Create a MapReduce Jobs workflow which framework can be used ?


 : To Create a MapReduce Jobs workflow which framework can be used ?
1. Sqoop
2. Hive
3. Pig
4. OOzie