Premium

Mapr (HP) Hadoop Developer Certification Questions and Answers (Dumps and Practice Questions)



Question : Which of the following method or methods of JobControl object can be used to track the execution state of Jobs
  : Which of the following method or methods of JobControl object can be used to track the execution state of Jobs
1. allFinished()
2. getFailedJobs()
3. Access Mostly Uused Products by 50000+ Subscribers
4. All of the above

Correct Answer : Get Lastest Questions and Answer :

JobControl object is used to control the group of Jobs, This class encapsulates a set of MapReduce jobs and its dependency
and there are various method get the status. It tracks the states of the jobs by placing them into different tables according to their states.
This class has a thread that submits jobs when they become ready, monitors the states of the running jobs, and updates the states of jobs based on the state changes of their
depending jobs states.
e.g. getFailedJobs(),getReadyJobs(),getRunningJobs() ,getState(),getSuccessfulJobs(),getWaitingJobs()




Question : Which class is use to preprocessing and postprocessing of a MapReduce Job
  : Which class is use to preprocessing and postprocessing of a MapReduce Job
1. ChainMapper
2. ChainReducer
3. Access Mostly Uused Products by 50000+ Subscribers
4. 1 and 2 Both

Correct Answer : Get Lastest Questions and Answer :

The ChainMapper class allows to use multiple Mapper classes within a single Map task.
The Mapper classes are invoked in a chained (or piped) fashion, the output of the first becomes the input of the second, and so on until the last Mapper, the output of the last
Mapper will be written to the task's output.

The ChainReducer class allows to chain multiple Mapper classes after a Reducer within the Reducer task.

For each record output by the Reducer, the Mapper classes are invoked in a chained (or piped) fashion, the output of the first becomes the input of the second, and so on until the
last Mapper, the output of the last Mapper will be written to the task's output.

Notes: Running all the Pre and Post processing in a single Jobs leaves no intermediate file and there is dramatic reduction in IO




Question : Is Data Joining like (RDBMS Join is possible in the Hadoop MapReduce)
  : Is Data Joining like (RDBMS Join is possible in the Hadoop MapReduce)
1. Yes
2. NO

Correct Answer : Get Lastest Questions and Answer :

There is a contrib package called datajoin that works as a generic framework for the data joining in Hadoop framework.


Related Questions


Question : Which is the correct command to delete the directory
  : Which is the correct command to delete the directory
1. hadoop fs -r pappu
2. hadoop fs -remove pappu
3. Access Mostly Uused Products by 50000+ Subscribers
4. hadoop fs -rem pappu


Question : Which statement is correct for the MapReduce
  : Which statement is correct for the MapReduce
1. MapReduce is a method for distribution a task across multiple nodes
2. Each node processes data stored on that node
3. Access Mostly Uused Products by 50000+ Subscribers
4. All of the above




Question : Which of the following are responsibilities of the ApplicationMater

1. Before starting any task, create job's output directory for job's OutputCommitter.
2. Both map tasks and reduce tasks are created by Application Master.
3. Access Mostly Uused Products by 50000+ Subscribers
4. If job doesn't qualify as Uber task, Application Master requests containers for all map tasks and reduce tasks.

 : Which of the following are responsibilities  of the ApplicationMater
1. 1,2,3
2. 2,3,4
3. Access Mostly Uused Products by 50000+ Subscribers
4. 1,2,4
5. 1,2,3,4





Question : A _____ is the basic unit of processing capacity in YARN, and is an encapsulation of resource elements (memory, cpu etc.)

 :  A _____ is the basic unit of processing capacity in YARN, and is an encapsulation of resource elements (memory, cpu etc.)
1. Node Manager
2. Container
3. Access Mostly Uused Products by 50000+ Subscribers
4. DataNode


Question : __________ are responsible for local monitoring of resource availability, fault reporting,
and container life-cycle management (e.g., starting and killing
jobs).


  : __________ are responsible for local monitoring of resource availability, fault reporting,
1. NodeManagers
2. Application Manager
3. Access Mostly Uused Products by 50000+ Subscribers
4. Resource Manager


Question : Your cluster has slave nodes in three different racks, and you have written a rack topology script identifying each machine as being in hadooprack,
hadooprack2, or hadooprack3. A client machine outside of the cluster writes a small (one-block) file to HDFS. The first replica of the block is written
to a node on hadooprack2. How is block placement determined for the other two replicas?
  : Your cluster has slave nodes in three different racks, and you have written a rack topology script identifying each machine as being in hadooprack,
1. One will be written to another node on hadooprack2, and the other to a node on a different rack.

2. Either both will be written to nodes on hadooprack1, or both will be written to nodes on hadooprack3.

3. Access Mostly Uused Products by 50000+ Subscribers

4. One will be written to hadooprack1, and one will be written to hadooprack3.