Premium

Mapr (HP) Hadoop Developer Certification Questions and Answers (Dumps and Practice Questions)



Question : In a cluster single node can run all the daemons ?
  : In a cluster single node can run all the daemons ?
1. Yes
2. No



Correct Answer : Get Lastest Questions and Answer :


Explanation: Technically it is possible to run all the hadoop daemons on single node, however it is not advisable

Refer HadoopExam.com Recorded Training Module : 2, 14 and 16





Question : MapR services produce metrics that can be written to an output file, which is the metrics that can be collected.
A. CLDB
B. File Servers
C. Hive Query Statistics
D. Fume job statistics
  : MapR services produce metrics that can be written to an output file, which is the metrics that can be collected.
1. A,B
2. B,C
3. Access Mostly Uused Products by 50000+ Subscribers
4. A,D
5. B,D

Correct Answer : Get Lastest Questions and Answer :
Explanation: By default, the CLDB and FileServer metrics are sent via unicast to the Ganglia gmon server running on localhost. To send the metrics directly to a Gmeta
server, change the
cldb.servers property to the hostname of the Gmeta server. To send the metrics to a multicast channel, change the cldb.servers property to the IP address of the multicast channel.



Question : MapReduce v (MRv/YARN) is designed to address which two issues?
A. Single point of failure in the NameNode.
B. Resource pressure on the JobTracker.
C. HDFS latency.
D. Ability to run frameworks other than MapReduce, such as MPI.
E. Reduce complexity of the MapReduce APIs.
F. Standardize on a single MapReduce API.

  : MapReduce v (MRv/YARN) is designed to address which two issues?
1. A,B
2. B,C
3. Access Mostly Uused Products by 50000+ Subscribers
4. A,D
5. B,D

Correct Answer : Get Lastest Questions and Answer :
Explanation:


Related Questions


Question : Which Daemon distributes individual task to machines
  : Which Daemon distributes individual task to machines
1. TaskTracker
2. JobTracker
3. Access Mostly Uused Products by 50000+ Subscribers
4. NameNode
Solution : 15



Question : You are using MapR Hadoop platform for Analyzing logs of WebServers you collected from Amazon Webservice.
However, you are not able to decide which one of the Pig or MapReduce you should use for this analysis. Now cpnsidering this scenario you have been asked to
find most accurately describes the relationship between MapReduce and Pig, Which opion you will see is more accurate?

 : You are using MapR Hadoop platform for Analyzing logs of WebServers you collected from Amazon Webservice.
1. Pig is more powerfull and allow certain types of data manipulation not possible with MapReduce.
2. Pig provides has same capabilities as MapReduce. Pig programs are executed as MapReduce jobs via the Pig interpreter.
3. Access Mostly Uused Products by 50000+ Subscribers
4. Pig provides the additional capability of allowing you to control the flow of multiple MapReduce jobs and Chaining the MapReduce jobs, which is not possible
only with the MapReduce.


Question :You have a WebSite called Quicktechie.com hosted on AWS (Amazon web service Cloud). You have been collecting your logs using Apache Flume and storing same in the
HDFS file system. However, while storing logs, you do the followings.
1. Remove all new line character from each log files (Size of individiual log file is not less than 200MB).
2. Append all the log files from same server together, however while appending log files you add new line character between two log files.
All the log files are in Text format and each newly created log file size (After appending) is not less than 100GB.
You selected TextInputFormat in your MapReduce job to read the logs data
for further processing. While log file processing, splits are representing data in HDFS. Select correct statement for split and new line character. ?

 :You have a WebSite called Quicktechie.com hosted on AWS (Amazon web service Cloud). You have been collecting your logs using Apache Flume and storing same in the
1. Input file splits may cross line breaks. A line that crosses tile splits is ignored.
2. The input file is split exactly at the line breaks, so each Record Reader will read a series of complete lines.
3. Access Mostly Uused Products by 50000+ Subscribers
4. Input file splits may cross line breaks. A line that crosses file splits is read by the RecordReader of the split that contains the end of the broken line.
5. Input file splits may cross line breaks. A line that crosses file splits is read by the RecordReader of the split that contains the beginning of the broken line.


Question : In a MapReduce job, you want each of you input files processed by a single map task.
How do you configure a MapReduce job so that a single map task processes each input
file regardless of how many blocks the input file occupies?


 : In a MapReduce job, you want each of you input files processed by a single map task.
1. Increase the parameter that controls minimum split size in the job configuration.
2. Write a custom MapRunner that iterates over all key-value pairs in the entire file.
3. Access Mostly Uused Products by 50000+ Subscribers
4. Write a custom FileInputFormat and override the method isSplittable to always return false.


Question : Given a directory of files with the following structure: line number, tab character, string: Example:
1. AvilakaurKohni
2. Decemberandaugust
3. Access Mostly Uused Products by 50000+ Subscribers
You want to send each line as one record to your Mapper.
Which InputFormat would you use to complete the line: setInputFormat (________.class);

 : Given a directory of files with the following structure: line number, tab character, string: Example:
1. BDBInputFormat
2. KeyValueTextInputFormat
3. Access Mostly Uused Products by 50000+ Subscribers
4. SequenceFileAsTextInputFormat


Question : What is a SequenceFile?

 : What is a SequenceFile?
1. A SequenceFile contains a binary encoding of an arbitrary number of homogeneous writable objects.
2. A SequenceFile contains a binary encoding of an arbitrary number of heterogeneous writable objects.
3. Access Mostly Uused Products by 50000+ Subscribers
4. A SequenceFile contains a binary encoding of an arbitrary number key-value pairs.
Each key must be the same type. Each value must be same type.