Premium

Cloudera Hadoop Developer Certification Questions and Answer (Dumps and Practice Questions)



Question : TaskTracker can not start multiple task in the same node

 :  TaskTracker can not start multiple task in the same node
1. True
2. False

Correct Answer : 2


Explanation: Submitting a Job
- When a client submits a job, its configuration information is packaged into an XML file.

This file, along with the .jar file containing the actual program code, is handed to the JobTracker
- The JobTracker then parcels out individual tasks to TaskTracker nodes
- When a TaskTracker receives a request to run a task, it instantiates a separate JVM for that task
- TaskTracker nodes can be configured to run multiple tasks at the same time
- If the node has enough processing power and memory

Refer HadoopExam.com Recorded Training Module : 3 and 4





Question : TaskTracker runs all the MapTask in the same JVM, if machine has enough processing power and Memory

 : TaskTracker runs all the MapTask in the same JVM, if machine has enough processing power and Memory
1. True
2. False

Correct Answer : 2

Submitting a Job
- When a client submits a job, its configuration information is packaged into an XML file.

This file, along with the .jar file containing the actual program code, is handed to the JobTracker
- The JobTracker then parcels out individual tasks to TaskTracker nodes
- When a TaskTracker receives a request to run a task, it instantiates a separate JVM for that task
- TaskTracker nodes can be configured to run multiple tasks at the same time
- If the node has enough processing power and memory

Refer HadoopExam.com Recorded Training Module : 3 and 4





Question : Select the correct statement


 : Select the correct statement
1. While job is running the intermediate data is keep deleted
2. Reducers write their final output to HDFS
3. Intermediate data is never deleted, HDFS stores them for History Tracking
4. All 1,2 and 3 are correct
5. None of the above


Correct Answers: 2

Explanation: Intermedate Data

The intermediate data is held on the TaskTrackers local disk
- As Reducers start up, the intermediate data is distributed across the network to the Reducers
- Reducers write their final output to HDFS
- Once the job has completed, the TaskTracker can delete the intermediate data from its local disk
- Note that the intermediate data is not deleted until the entire job completes

Refer HadoopExam.com Recorded Training Module : 2,3 and 4



Related Questions


Question :

Which language is used in oozie to define MapReduce workflow
 :
1. Java
2. XML
3. Pig Latin
4. None of the above




Question :

Select the correct statement?
 :
1. In oozie workflow, all the MapReduce jobs can run in sequence only
2. Jobs can run parallel as well as in sequence
3. One Job can wait for another job to finish
4. All of the above
5. 2 and 3



Question :

What is the benefit of the Hadoop framework
 :
1. Can process PetaBytes of data volume
2. Can store PetaBytes of Structured and Unstructured data
3. data can be imported and exported from RDBMS
4. All of the above




Question :

In which of the below scenerio Hadoop framework should not be used
 :
1. Working with small and structured datasets
2. If there is a requirement of transaction
3. For Structured analytics
4. 1 and 2
5. 1 , 2 and 3


Question :

DistCp tool is used for
 :
1. Tool used for large inter or intra cluster copying
2. Uses MapReduce for distribution
3. 1 and 2 are correct
4. None of the above




Question :

Which of the following data format can be analyzed by Hadoop

 :
1. XML
2. CSV
3. JSON
4. Text
5. All of the above