Premium

Cloudera HBase Certification Questions and Answers (Dumps and Practice Questions)



Question :

While inserting the data to HBase using Put

 :
1. True
2. False




Correct Answer : Get Lastest Questions and Answer :

While Distributing the load, problem is Sequential keys such as timestamps will go to the same region
to solve this issue Prefix key with a hash or other well distributed value for better throughput of writes







Question :

While schema design which of the following is valid point to keeping StoreFile indices small..



 :
1. Keep ColumnFamily names as small as possible
2. Avoid long verbose attribute names
3. Access Mostly Uused Products by 50000+ Subscribers
4. All 1,2 and 3 are correct
5. Only 1 and 3 are correct



Correct Answer : Get Lastest Questions and Answer :

While Schema Design Large StoreFile indices
- Every cell always includes row, column name and timestamp
- Indices are kept in HBase StoreFiles to facilitate random access
- Large cell value coordinates increase the size of indices
- May occupy large chunks of RAM
- Compression also increases the size of indices
Increase the block size
- Store file index will happened at a larger interval
Keep names small
- Keep ColumnFamily names as small as possible
- Avoid long verbose attribute names
- Keep RowKey length as short as is reasonable





Question :

For storing the data in HBase
Anything that can be converted to an array of bytes can be stored
  :
1. True
2. False



Correct Answer : Get Lastest Questions and Answer :

Bytes-in/bytes-out interface
Anything that can be converted to an array of bytes can be stored
- Input can be strings, numbers, complex objects, images, etc.





Related Questions


Question : Rows from the HBase can directly be inserted as input to Mapreduce job

 :  Rows from the HBase can directly be inserted as input to Mapreduce job
1. True
2. False



Question : In which of the following scenerio we should use HBase
 :  In which of the following scenerio we should use HBase
1. If it require random read, write or both
2. If it requires to do many thousands of operations per second on multiple TB of data
3. If access pattern is well known and simple
4. All of the above




Question : In which scenerio HBase should not be used

 :  In which scenerio HBase should not be used
1. You only append to your dataset, and tend to read the whole thing
2. For ad-hoc analytics
3. If data volume is quite small
4. All of the above
5. None of the above



Question : Which is the correct statemnet to inserting the data in table

t1 and row key r1 and column family fam1

 :  Which is the correct statemnet to inserting the data in table
1. put 't1', 'r1', 'fam1:c1', 'value'
2. put 't1', 'r1', 'fam1:c1', 'value', 1274302629663
3. 1 and 2 are correct


Question : Which is the wrong syntex for scan

 :  Which is the wrong syntex for scan
1. scan 't1'
2. scan 't1', {COLUMNS => 'fam1:c1'}
3. scan 't1', {COLUMNS => 'fam1:'}
4. scan 't1', {STARTROW => 'r1', LIMIT => 10}
5. scan 't1', {COLUMNS => ':c1'}



Question :

Which of the following is a wrong statement abount count

operation on HBase table
 :
1. Syntex for count is count 'tablename' [, interval]
2. Counting the rows of a large table can be slow
3. The progress will be reported every interval
4. None of the 1,2 and 3
5. Only 1 and 2