Premium

Mapr (HP) Hadoop Developer Certification Questions and Answers (Dumps and Practice Questions)



Question : Select the correct flow by which MapReduce job process the data in MRv.

  : Select the correct flow by which MapReduce job process the data in MRv.
1. Input File > Mapper > Combiner (Optional) > shuffle > reducer >output file

2. Mapper > Input File > Combiner (Optional) > shuffle > reducer >output file

3. Input File > Mapper > shuffle > Combiner (Optional) > reducer >output file

4. Input File > Mapper > shuffle > reducer > Combiner (Optional) >output file

Correct Answer : 1
Explanation:




Question : If you have written a Mapper which has both Input key and Value as Text. Which of the following component you will us eto break entire text in the Individual fields.

  : If you have written a Mapper which has both Input key and Value as Text. Which  of the following component you will us eto break entire text in the Individual fields.
1. NameNode

2. TaskTracker

3. Application Master

4. StringTokenizer

Correct Answer : 4
Explanation:




Question : As you have following map() function of the Mapper

public void map(Text text, Text value, Content cnt) throws ....
cnt.write(new Text("Hadoop"), new Text("Exam"));

Whar is the new Text("Hadoop")


  : As you have following map() function of the Mapper
1. Key

2. Value

3. Represent to use MRv1 framework

4. Represent to use MRv2 framework

Correct Answer : 1
Explanation:


Related Questions


Question : In Hadoop ., which TWO of the following processes work together to provide automatic
failover of the NameNode? Choose 2 answers
A. ZKFailoverController
B. ZooKeeper
C. QuorumManager
D. JournalNode

  : In Hadoop ., which TWO of the following processes work together to provide automatic
1. A,B
2. B,C
3. Access Mostly Uused Products by 50000+ Subscribers
4. A,D
5. B,D


Question : Which one of the following statements regarding the components of YARN is FALSE?
  : Which one of the following statements regarding the components of YARN is FALSE?
1. A Container executes a specific task as assigned by the ApplicationMaster
2. The ResourceManager is responsible for scheduling and allocating resources
3. Access Mostly Uused Products by 50000+ Subscribers
4. The ResourceManager monitors and restarts any failed Containers


Question : Analyze each scenario below and indentify which best describes the behavior of the default partitioner?

  : Analyze each scenario below and indentify which best describes the behavior of the default partitioner?
1. The default partitioner assigns key-values pairs to reduces based on an internal random number generator.
2. The default partitioner implements a round-robin strategy, shuffling the key-value pairs
to each reducer in turn. This ensures an event partition of the key space.
3. Access Mostly Uused Products by 50000+ Subscribers
ranges are associated with different buckets, and each bucket is assigned to a specific reducer.
4. The default partitioner computes the hash of the key and divides that value modulo the
number of reducers. The result determines the reducer assigned to process the key-value pair.
5. The default partitioner computes the hash of the value and takes the mod of that value
with the number of reducers. The result determines the reducer assigned to process the key-value pair.


Question : In a MapReduce job, the reducer receives all values associated with same key. Which
statement best describes the ordering of these values?
  : In a MapReduce job, the reducer receives all values associated with same key. Which
1. The values are in sorted order.
2. The values are arbitrarily ordered, and the ordering may vary from run to run of the same MapReduce job.
3. Access Mostly Uused Products by 50000+ Subscribers
4. Since the values come from mapper outputs, the reducers will receive contiguous sections of sorted values.


Question : You have just executed a MapReduce job. Where is intermediate data written to after being emitted from the Mapper's map method?

  : You have just executed a MapReduce job. Where is intermediate data written to after being emitted from the Mapper's map method?
1. Intermediate data in streamed across the network from Mapper to the Reduce and is never written to disk.
2. Into in-memory buffers on the TaskTracker node running the Mapper that spill over and are written into HDFS.
3. Access Mostly Uused Products by 50000+ Subscribers
4. Into in-memory buffers that spill over to the local file system (outside HDFS) of the TaskTracker node running the Reducer
5. Into in-memory buffers on the TaskTracker node running the Reducer that spill over and are written into HDFS.


Question : You are developing a MapReduce job for sales reporting. The mapper will process input
keys representing the year (IntWritable) and input values representing product identifies
(Text). Identify what determines the data types used by the Mapper for a given job.

  : You are developing a MapReduce job for sales reporting. The mapper will process input
1. The key and value types specified in the JobConf.setMapInputKeyClass and JobConf.setMapInputValuesClass methods
2. The data types specified in HADOOP_MAP_DATATYPES environment variable
3. Access Mostly Uused Products by 50000+ Subscribers
4. The InputFormat used by the job determines the mapper's input key and value types.