Premium

Mapr (HP) Hadoop Developer Certification Questions and Answers (Dumps and Practice Questions)



Question : To check the MapR Job Performance, we use MapR Control System. However, a job completed with % map task and % reduce tasks and Job is not finishing.
So you can use MapR control system as
A. You can filter the views in the MapR Control System to list only reduce tasks
B. Once you have a list of your job's reduce tasks, you can sort the list by duration to see if any reduce task attempts are taking an abnormally long time to execute
C. you can not filter the views in the MapR Control System to list only reduce tasks
D. MapR Control System can display detailed information about those task attempts, including log files for those task attempts
 : To check the MapR Job Performance, we use MapR Control System. However, a job completed with % map task and % reduce tasks and Job is not finishing.
1. A,B,C
2. B,C,D
3. Access Mostly Uused Products by 50000+ Subscribers
4. A,B,D

Correct Answer : Get Lastest Questions and Answer :
Explanation: if a job lists 100% map task completion and 99% reduce task completion, you can filter the views in the MapR Control System to list only reduce tasks. Once
you have a list of
your job's reduce tasks, you can sort the list by duration to see if any reduce task attempts are taking an abnormally long time to execute, then display detailed information about
those task attempts, including log files for those task attempts.




Question : Can we use MapR control system Metrics displays to gauge performance of two different jobs that perform the same function one written n Python using pydoop and other is
written in C++ using Pipes
 : Can we use MapR control system Metrics displays to gauge performance of two different jobs that perform the same function one written  n Python using pydoop and other is
1. Yes
2. No

Correct Answer : Get Lastest Questions and Answer :
Explanation: You can also use the Metrics displays to gauge performance. Consider two different jobs that perform the same function. One job is written in Python using
pydoop,
and the other job is written in C++ using Pipes. To evaluate how these jobs perform on the cluster, you can open two browser windows logged into the MapR Control System and filter
the display down to the metrics you're most interested in while the jobs are running.





Question : To use MapR Metrics, set up a ________ database to log metrics data.

 : To use MapR Metrics, set up a ________ database to log metrics data.
1. MySQL

2. Oracle

3. Access Mostly Uused Products by 50000+ Subscribers

4. SQL Server

Correct Answer : Get Lastest Questions and Answer :
Explanation: To use MapR Metrics, set up a MySQL database to log metrics data. The MapR distribution for Apache Hadoop does not include MySQL. Download and install MySQL
separately and then perform the configuration steps to enable the MapR Metrics database.




Related Questions


Question : You've written a MapReduce job based on HadoopExam websites log file named MAIN.PROFILE.log file , resulting in an extremely
large amount of output data. Which of the following cluster resources will your job stress? ?
 : You've written a MapReduce job based on HadoopExam websites log file named MAIN.PROFILE.log file , resulting in an extremely
1. network I/O and disk I/O
2. network I/O and RAM
3. Access Mostly Uused Products by 50000+ Subscribers
4. RAM , network I/O and disk I/O


Question : You have written a Mapper which invokes the following five calls to the OutputCollector.collect method:

output.collect(new Text("Flag"), new Text("Rahul"));
output.collect(new Text("Shirt"), new Text("Yakul"));
output.collect(new Text("Shoe"), new Text("Rahul"));
output.collect(new Text("Flag"), new Text("Gemini"));
output.collect(new Text("Socks"), new Text("Yakul"));

How many times will the Reducer's reduce() method be invoked.

 : You have written a Mapper which invokes the following five calls to the OutputCollector.collect method:
1. 5
2. 4
3. Access Mostly Uused Products by 50000+ Subscribers
4. 7
5. 8


Question : ___________ is an optimization technique where a computer system performs some task that may not be actually needed. The main idea is to
do work before it is known whether that work will be needed at all, so as to prevent a delay that would have to be incurred by doing the work after it
is known whether it is needed. If it turns out the work was not needed after all, the results are ignored. The Hadoop framework also provides a
mechanism to handle machine issues such as faulty configuration or hardware failure. The JobTracker detects that one or a number of
machines are performing poorly and starts more copies of a map or reduce task. This behaviour is known as ________________

 : ___________ is an optimization technique where a computer system performs some task that may not be actually needed. The main idea is to
1. Task Execution
2. Job Execution
3. Access Mostly Uused Products by 50000+ Subscribers
4. Speculative Execution


Question :
You are working in the HadoopExam consultancy team and written a MapReduce and Pig job, which of the following is correct statement?

  :
1. Pig comes with additional capabilities to MapReduce. Pig programs are executed as MapReduce jobs via the Pig interpreter.
2. Pig comes with no additional capabilities to MapReduce. Pig programs are executed as MapReduce jobs via the Pig interpreter.
3. Access Mostly Uused Products by 50000+ Subscribers
4. Pig comes with additional capabilities to MapReduce. Pig programs are executed as MapReduce jobs via the Pig interpreter.


Question : Everyday HadoopExam has a good number of subscribers, but the file size created from this information is
smaller than 64MB, and same 64MB is configured as a block size on the cluster.
You are running a job that will process this file as a single input split on a cluster which has no other jobs currently running,
and with all settings at their default values. Each node has an equal number of open Map slots.
On which node will Hadoop first attempt to run the Map task?

  : Everyday HadoopExam has a good number of subscribers, but the file size created from this information is
1. The node containing the first TaskTracker to heartbeat into the JobTracker, regardless of the location of the input split
2. The node containing the first JobTracker to heartbeat into the Namenode, regardless of the location of the input split
3. Access Mostly Uused Products by 50000+ Subscribers
4. The node containing nearest location of the input split


Question : You are working on a project of HadoopExam client where you need to chain together MapReduce and Pig jobs.
You also need the ability to use forks, decision points, and path joins.
Which of the following ecosystem projects allows you to accomplish this?

 : You are working on a project of HadoopExam client where you need to chain together MapReduce and Pig jobs.
1. Oozie
2. MapReduce chaining
3. Access Mostly Uused Products by 50000+ Subscribers
4. Zookeeper
5. Hue