Question : TaskTracker can not start multiple task in the same node
1. True 2. False
Correct Answer : 2
Explanation: Submitting a Job - When a client submits a job, its configuration information is packaged into an XML file.
This file, along with the .jar file containing the actual program code, is handed to the JobTracker - The JobTracker then parcels out individual tasks to TaskTracker nodes - When a TaskTracker receives a request to run a task, it instantiates a separate JVM for that task - TaskTracker nodes can be configured to run multiple tasks at the same time - If the node has enough processing power and memory
Refer HadoopExam.com Recorded Training Module : 3 and 4
Question : TaskTracker runs all the MapTask in the same JVM, if machine has enough processing power and Memory
1. True 2. False
Correct Answer : 2
Submitting a Job - When a client submits a job, its configuration information is packaged into an XML file.
This file, along with the .jar file containing the actual program code, is handed to the JobTracker - The JobTracker then parcels out individual tasks to TaskTracker nodes - When a TaskTracker receives a request to run a task, it instantiates a separate JVM for that task - TaskTracker nodes can be configured to run multiple tasks at the same time - If the node has enough processing power and memory
Refer HadoopExam.com Recorded Training Module : 3 and 4
Question : Select the correct statement
1. While job is running the intermediate data is keep deleted 2. Reducers write their final output to HDFS 3. Intermediate data is never deleted, HDFS stores them for History Tracking 4. All 1,2 and 3 are correct 5. None of the above
Correct Answers: 2
Explanation: Intermedate Data
The intermediate data is held on the TaskTrackers local disk - As Reducers start up, the intermediate data is distributed across the network to the Reducers - Reducers write their final output to HDFS - Once the job has completed, the TaskTracker can delete the intermediate data from its local disk - Note that the intermediate data is not deleted until the entire job completes
Refer HadoopExam.com Recorded Training Module : 2,3 and 4
Select the correct statement? 1. In oozie workflow, all the MapReduce jobs can run in sequence only 2. Jobs can run parallel as well as in sequence 3. One Job can wait for another job to finish 4. All of the above 5. 2 and 3
What is the benefit of the Hadoop framework 1. Can process PetaBytes of data volume 2. Can store PetaBytes of Structured and Unstructured data 3. data can be imported and exported from RDBMS 4. All of the above
DistCp tool is used for 1. Tool used for large inter or intra cluster copying 2. Uses MapReduce for distribution 3. 1 and 2 are correct 4. None of the above