Premium

Mapr (HP) HBase Developer Certification Questions and Answers (Dumps and Practice Questions)



Question : Because MapR-DB does not require ______________ to be reconstructed, databases can be brought up instantly on the mirrored site if the active site goes down.

 : Because MapR-DB does not require ______________ to be reconstructed, databases can be brought up instantly on the mirrored site if the active site goes down.
1. Volume

2. RegionServers

3. Access Mostly Uused Products by 50000+ Subscribers

4. Tables

Correct Answer : Get Lastest Questions and Answer :
Explanation: Mirroring of volumes lets you automatically replicate differential data in real-time across clusters. You might want to do this to create disaster recovery
solutions for databases or to provide read-only access to data from multiple locations. Because MapR-DB does not require RegionServers to be reconstructed, databases can be brought
up instantly on the mirrored site if the active site goes down.





Question : During data mirroing, MapR copies the entire file block, if it has changes since last time mirroring.

 : During data mirroing, MapR copies the entire file block, if it has changes since last time mirroring.
1. True
2. False

Correct Answer : Get Lastest Questions and Answer :
Explanation: Mirroring of volumes lets you automatically replicate differential data in real-time across clusters. You might want to do this to create disaster recovery
solutions for databases or to provide read-only access to data from multiple locations. Because MapR-DB does not require RegionServers to be reconstructed, databases can be brought
up instantly on the mirrored site if the active site goes down.
Mirroring is a parallel operation, copying data directly from the nodes of one MapR cluster to the nodes in a remote MapR cluster. The contents of the volume are mirrored
consistently, even if the files in the volume are being written to or deleted.
MapR captures only data that has changed at the file-block level since the last data transfer. After the data differential is identified, it is then compressed and transferred
over the WAN to the recovery site, using very low network bandwidth. Finally, checksums are used to ensure data integrity across the two clusters. There is no performance penalty
on the cluster because of mirroring.





Question : MapR-DB is a multi-model database, supporting the easy-to-use

 : MapR-DB is a multi-model database, supporting the easy-to-use
1. A,B
2. B,C
3. Access Mostly Uused Products by 50000+ Subscribers
4. A,D
5. B,D

Correct Answer : Get Lastest Questions and Answer :
Explanation: MapR-DB is an enterprise-grade, high performance, in-Hadoop NoSQL ("Not Only SQL") database management system. It is used to add real-time, operational
analytics capabilities to Hadoop. NoSQL primarily addresses two critical data architecture requirements:

Scalability to address the increasing volumes and velocity of your data
Flexibility to store the variety of useful data types and formats
MapR-DB goes beyond other NoSQL solutions by providing a reliable platform to run business-critical big data applications. MapR-DB is a multi-model database, supporting the
easy-to-use, open source OJAIā„¢(Open JSON Application Interface) API for document database capabilities, as well as an API for wide column data models.



Related Questions


Question : There is a feature provided in QuickTechie.com website that any Software Professional can create an article as well as can update and delete the
article. You decided to use HBase rather than HDFS to store this article. You have created an ARTICLES table in HBase to store all the versions of the articles
in this table. Select the Column Family attribute settings which will retain at least one version of an article always but expire all other versions that are older than
1 month (30 Days) for a given Column Family?
  :  There is a feature provided in QuickTechie.com website that any Software Professional can create an article as well as can update and delete the
1. LENGTH = 30, MIN_VERSIONS = 1
2. TTL = 30, VERSIONS = 1

3. Access Mostly Uused Products by 50000+ Subscribers

4. TTL = 2592000 , MIN_VERSIONS = 1





Question : Given that following is your entire data set:

23 column=Engineers:FirstName, timestamp=1331314762084, value=Shobhit
23 column=Engineers:Payment, timestamp=1331314762086, value=800000
23 column=TechnicalSkills:1_FirstSkill, timestamp=1331314762089, value=J2EE
23 column=TechnicalSkills:2_AnotherSkill, timestamp=1331314762092, value=Java

How many sets of physical files will be read during a scan of the entire data set immediately following a major compaction?
  :   Given that following is your entire data set:
1. One
2. Two
3. Access Mostly Uused Products by 50000+ Subscribers
4. Four




Question : As an HBase administrator at Acmeshell.com you have configured HBase to store a maximum of versions.
You have inserted 7 versions of your data in a Column Family called Acmeshell. At what point are the older versions removed from Acmeshell?
  :  As an HBase administrator at Acmeshell.com you have configured HBase to store a maximum of  versions.
1. Never, the older version has to be manually deleted.
2. The older versions are removed at major compaction.
3. Access Mostly Uused Products by 50000+ Subscribers
4. The older versions are removed at minor compaction.




Question : To analysing the entire QuickTechie.com articles backup table stored in HBase, you found that it is not perfoming well and showing slowness.
You considered the block size option and increased the block size from 64KB to 512KB assuming ARTICLE table size is 1TB. Why does increasing block size improve scan performance?
  :   To analysing the entire QuickTechie.com articles backup table stored in HBase, you found that it is not perfoming well and showing slowness.
1. When you increase block size then HBase will reduce the seek on the disk by which scan performance increased.
2. Increasing block size means fewer block indexes that need to be read from disk, which increase scan performance.
3. Access Mostly Uused Products by 50000+ Subscribers
4. None of the above




Question : You have written a Mobile Application for an advertising company called Acmeshell.com.
Your Mobile application needs to retrieve 01011990(start date) to 31121990 (end date) non-sequential rows from a table with 1TB mobiles ads stored as rows .
What does your application need to implement to retrieve data for known row keys ?
  :  You have written a Mobile Application for an advertising company called Acmeshell.com.
1. HTable.get(List(Get) gets)

2. Increase the Block Cache
3. Access Mostly Uused Products by 50000+ Subscribers
4. HTable.get(Get get)





Question : You have created an HBase application called Acmeshell and from within Acmeshell, you want to create a new table named AcmeLogs.
In this AcmeLogs table you will be storing 2 Billion Mobile advertisement and its clickstream information. You start with the following Java code:
You have already created HBaseAdmin object ( name of object acmeAdmin)using the configuration as well as HTableDescriptor with
table name "AcmeLogs". Now you want to finally create the table using HTableDescriptor, select the correct command.
  :  You have created an HBase application called Acmeshell and from within Acmeshell, you want to create a new table named AcmeLogs.
1. HTable.createTable(acmeTable);
2. HBaseAdmin.createTable(acmeTable);
3. Access Mostly Uused Products by 50000+ Subscribers

4. acmeAdmin.createTable(acmeTable);