Premium

Mapr (HP) Hadoop Developer Certification Questions and Answers (Dumps and Practice Questions)



Question : Select correct statement regarding SequenceFile
A. '\n' is used as a record terminator
B. Sync marker is used as a record terminator
C. The key-value records are bundled into blocks.
D. The block delimiters are called "markers", and the size of a block is tunable
 : Select correct statement regarding SequenceFile
1. A,C,D
2. A,B,D
3. Access Mostly Uused Products by 50000+ Subscribers
4. A,B,C
5. A,B,C,D

Correct Answer : Get Lastest Questions and Answer :
Explanation:




Question : You have to write a Job , which reads SequenceFile and produce as an output Compressed SequenceFile (Compression type is Gzip). Below is the
code snippet for Driver class

setOutputFormatClass(1)
setCompressOutput(2)
setOutputCompressorClass(3)
setOutputCompressionType(4)
setInputFormatClass(5)

Map the below.

B. job,true
C. job, GzipCodec.class
D. job, CompressionType.BLOCK
E. SequenceFileInputFormat.class
A. SequenceFileOutputFormat.class

 : You have to write a Job , which reads SequenceFile and produce as an output Compressed SequenceFile (Compression type is Gzip). Below is the
1. 1-B, 2-A, 3-D, 4-C, 5-E
2. 1-E, 2-B, 3-C, 4-D, 5-A
3. Access Mostly Uused Products by 50000+ Subscribers
4. 1-A, 2-B, 3-C, 4-D, 5-E
5. 1-D, 2-B, 3-C, 4-A, 5-E

Correct Answer : Get Lastest Questions and Answer :
Explanation:




Question : Using SequenceFile can save disk space as well as time if more than one MapReduce jobs are chained together
 : Using SequenceFile can save disk space as well as time if more than one MapReduce jobs are chained together
1. True
2. False

Correct Answer : Get Lastest Questions and Answer :
Explanation:


Related Questions


Question : Which is a true statement regarding MapR output compressions and codec?


 : Which is a true statement regarding MapR output compressions and codec?
1. Default compression codec in MapR is LZ4

2. ZLIB codec provides the highest level of compression but also requires the most CPU cycles.

3. Access Mostly Uused Products by 50000+ Subscribers

4. The data is compressed which is spilled to local disk in sort phase on the Mappers

5. Configure LD_LIBRARY_PATH to enable native codecs



Question : You have written a MapReduce job, which process months data in Mapper for year . However, all the required transformation is done in
Mapper only.

And you just want output to be stored in specified directory. How many reduce tasks you should configure


 : You have written a MapReduce job, which process  months data in Mapper for year . However, all the required transformation is done in
1. You cannot decide number of reducer. It is decided by Framework only

2. 0

3. Access Mostly Uused Products by 50000+ Subscribers

4. 1



Question : Which of the following property will help, defining number of reducer tasks on job level


 : Which of the following property will help, defining number of reducer tasks on job level
1. mapred.reduce.tasks

2. mapred.max.reduce.tasks

3. Access Mostly Uused Products by 50000+ Subscribers

4. mapred.min.reduce.tasks



Question : Which of the following property will help to limit number of reduce tasks across all the Jobs


 : Which of the following property will help to limit number of reduce tasks across all the Jobs
1. mapred.reduce.tasks

2. mapred.max.reduce.tasks

3. Access Mostly Uused Products by 50000+ Subscribers

4. mapred.min.reduce.tasks



Question : What is the ideal range for mapred.reduce.tasks parameters?


 : What is the ideal range for mapred.reduce.tasks parameters?
1. between 0.95 to 1.75

2. between 1 to 2

3. Access Mostly Uused Products by 50000+ Subscribers

4. You can choose any range, and it will derived based on relative value.



Question : You have a MapReduce jobs, which create unique data sets and finally insert each record in JDBC database table. Reducer is responsible writing
data in Database. There are chances that your cluster is very heavily loaded and few map tasks and reduce tasks can fail in between and re-launched in
different node. So, which statement is correct for above scenario?



 : You have a MapReduce jobs, which create unique data sets and finally insert each record in JDBC database table. Reducer is responsible writing
1. to avoid slowness, we should enable speculative execution

2. we should not enable speculative execution

3. Access Mostly Uused Products by 50000+ Subscribers

4. We should reduce number of reducer