Premium

Cloudera Hadoop Developer Certification Questions and Answer (Dumps and Practice Questions)



Question : Select the correct statement ?
 : Select the correct statement ?
1. Block size is usually 64 MB or 128 MB
2. Blocks are replicated across multiple machine
3. Access Mostly Uused Products by 50000+ Subscribers
4. All of the above

Correct Answer : Get Lastest Questions and Answer :





Question : Which is the master node for tracking the files block in HDFS ?

 : Which is the master node for tracking the files block in HDFS ?
1. JOBTracker
2. DataNode
3. Access Mostly Uused Products by 50000+ Subscribers
4. DataMasteNode

Correct Answer : Get Lastest Questions and Answer :

A master node called the Namenode which keeps track of all the blocks in HDFS which make up
a comlete file, and where those blocks are located. By default the replication factor in the HDFS
is three. Different blocks from the same file will be stored on different machines.




Question : Select the correct options

 : Select the correct options
1. NameNode store the metadata for the files
2. DataNode holds the actual blocks
3. Access Mostly Uused Products by 50000+ Subscribers
4. All of the above
5. 1 and 2 are correct

Correct Answer : Get Lastest Questions and Answer :




Related Questions


Question :

You have defined flume agent as a1, with following configuration

a1.channels = c1
a1.sinks = k1
a1.sinks.k1.type = hdfs
a1.sinks.k1.channel = c1
a1.sinks.k1.hdfs.path = /flume/events/%y-%m-%d/%H%M/%S
a1.sinks.k1.hdfs.filePrefix = events-
a1.sinks.k1.hdfs.round = true
a1.sinks.k1.hdfs.roundValue = 10
a1.sinks.k1.hdfs.roundUnit = minute

an event with timestamp 11:54:34 AM, June 12, 2012 will cause the hdfs path to become

  :
1. /flume/events/2012-06-12/1150/00
2. /flume/events/2012-06-12/1200/00
3. Access Mostly Uused Products by 50000+ Subscribers
4. /flume/events/2012-06-12/1160/00


Question :

You have defined flume agent as a1, with following configuration

a1.channels = c1
a1.sinks = k1
a1.sinks.k1.type = hdfs
a1.sinks.k1.channel = c1
a1.sinks.k1.hdfs.path = /flume/events/%y-%m-%d/%H%M/%S
a1.sinks.k1.hdfs.filePrefix = events-
a1.sinks.k1.hdfs.round = true
a1.sinks.k1.hdfs.roundValue = 20
a1.sinks.k1.hdfs.roundUnit = minute

an event with timestamp 11:51:34 AM, June 12, 2012 and another event happens as 11:54:34 AM, June 12, 2012
So in which of the path the file will be stored

 :
1. /flume/events/2012-06-12/1140/00
2. /flume/events/2012-06-12/1200/00
3. Access Mostly Uused Products by 50000+ Subscribers
4. /flume/events/2012-06-12/1160/00


Question :

There are two input files as belwo to MapReduce Join job.

input/A
A.a11 A.a12
A.a21 A.a22
B.a21 A.a32
A.a31 A.a32
B.a31 A.a32

input/B
A.a11 B.a12
A.a11 B.a13
B.a11 B.a12
B.a21 B.a22
A.a31 B.a32
B.a31 B.a32

After running the MapReduce join code snippet(Left Hand Side)

What would be the first line of the output

 :
1. A.a11 A.a12 B.a12
2. A.a11 A.a12 A.a11 B.a13
3. Access Mostly Uused Products by 50000+ Subscribers
4. B.a21 A.a32 B.a21 B.a22


Question :

Select the correct code snippet which will produce the 12 files each for a month, considering you have defined 12 reducres for this job

Sample input data
10.1.255.266,hadoopexam.com,index.html,20/Aug/2013
10.1.255.2,hadoopexam.com,index.html,11/Feb/2013
10.1.255.233,hadoopexam.com,index.html,14/Jan/2013

 :
1. 1
2. 2
3. Access Mostly Uused Products by 50000+ Subscribers


Question :

From the below given code snippet please select the correct one which is able to create Compressed Sequence file.

 :
1. 1
2. 2
3. Access Mostly Uused Products by 50000+ Subscribers


Question :

Select the correct code snippet which is able to read the compressed sequence file

 :
1. 1
2. 2
3. Access Mostly Uused Products by 50000+ Subscribers