Premium
Cloudera Hadoop Developer Certification Questions and Answer (Dumps and Practice Questions)
Question :
Which statement is correct for the MapReduce
1. MapReduce is a method for distribution a task across multiple nodes
2. Each node processes data stored on that node
3. Access Mostly Uused Products by 50000+ Subscribers
4. All of the above
Correct Answer
:
Get Lastest Questions and Answer
:
Question :
Select the feature of Mapreduce
1. Automatic parallelization and distribution
2. fault-tolerance
3. Access Mostly Uused Products by 50000+ Subscribers
4. All of the above
Correct Answer
:
Get Lastest Questions and Answer
:
Question : Hadoop Mapreduce code can be written in other than Java language ?
1. True
2. False
Correct Answer
:
Get Lastest Questions and Answer
:
Mapreduce programs can be written in any scripting language using Hadoop Streaming
Related Questions
Question : You have created a MapReduce job to process TimeSeries Market Data file with the driver class called
HadoopDriver (in the default package) packaged into a jar called HadoopExam.jar, what is the appropriate way to submit this job to the cluster?
1. hadoop jar HadoopExam.jar HadoopDriver outputdir inputdir
2. hadoop inputdir outputdir jar HadoopExam.jar HadoopDriver
3. Access Mostly Uused Products by 50000+ Subscribers
4. hadoop jar HadoopExam.jar HadoopDriver inputdir outputdir
Question : To analyze the website click of HadoopExam.com you have written a Mapreduce job, which
will product the click reports for each week e.g. 53 reports for whole year.Which of the following Hadoop API class you must use
so that output file generated as per the weeks and output data will go in corresponding output file.
1. Hive
2. MapReduce Chaining
3. Access Mostly Uused Products by 50000+ Subscribers
4. Partitioner
Question : Reducers are generally helpful to write the job ouput data in desried location or database.
In your ETL MapReduce job you set the number of reducer to zero, select the correct statement which applies.
1. You can not configure number of reducer
2. No reduce tasks execute. The output of each map task is written to a separate file in HDFS
3. Access Mostly Uused Products by 50000+ Subscribers
4. You can not configure number of reducer, it is decided by Tasktracker at runtime
Question : In the QuickTechie website log file named as MAIN.PROFILES.log you have keys are (ipaddres+locations), and the values are Number of clicks (int).
For each unique key (string), you want to find the average of all values associated with each key. In writing a MapReduce program to accomplish this, can you take advantage of a combiner?
1. No, best way to accomplish this you have to use Aapche Pig
2. No, best way to accomplish this you have to use MapReduce chaining.
3. Access Mostly Uused Products by 50000+ Subscribers
4. Yes
Question : In our website www.HadoopExam.com we have Million profiles and created ETL jobs for procesing this file.
You have submited a ETL mapReduce job for HadoopExam.com websites log file analysis as well as combining profile data to Hadoop
and notice in the JobTracker's Web UI that the Mappers are 80% complete
while the reducers are 20% complete. What is the best explanation for this?
1. The progress attributed to the reducer refers to the transfer of data from completed Mappers.
2. The progress attributed to the reducer refers to the transfer of data from Mappers is still going on.
3. Access Mostly Uused Products by 50000+ Subscribers
4. The progress attributed to the reducer refers to the transfer of data from Mappers an not be predicted.
Question : In your MapReduce job, you have three configuration parameters.
What is the correct or best way to pass a these three configuration parameters to a mapper or reducer?
1. As key pairs in the Configuration object.
2. As value pairs in the Configuration object.
3. Access Mostly Uused Products by 50000+ Subscribers
4. Not possible