Premium

IBM Certified Data Architect - Big Data Certification Questions and Answers (Dumps and Practice Questions)



Question : You are working in a company known as Arinika INC., which has partnership with IBM and your are providing consultancy services for data analytics for a financial
institution. This institute receives 1 TB data per day. You have been asked that financial institutes data scientist should be able to run fast data analytics after feeding the
data in system. Which of the following is best tool provided by IBM to support such problem?

  : You are working in a company known as Arinika INC., which has partnership with IBM and your are providing consultancy services for data analytics for a financial
1. IBM Reporting Tool

2. IBM lotus

3. Access Mostly Uused Products by 50000+ Subscribers

4. Inforsphere Streams

5. IBM PureData System

Correct Answer : Get Lastest Questions and Answer :
Explanation: IBM PureData System for Analytics is a purpose-built, standards-based data warehouse and analytics appliance that integrates database, server, storage and
analytics into an easy-to-manage system. It is designed for high-speed analysis of big data volumes, scaling into the petabytes.




Question : Your have team of data scientist and you are facing lot of problems, while working with high volume of data. Hence, one of the solution architect in your company
suggested to use IBM BIgSQL, which statement is true for BigSQL


A. Big SQL, IBM's SQL interface to its Hadoop-based platform, InfoSphere BigInsights.
B. It enables data administrators to create new tables for data stored in Hive, HBase, or their BigInsights distributed file system.
C. Big SQL can use Hadoop's MapReduce framework to process various query tasks in parallel or execute your query locally within the Big SQL server on a single node
D. Big SQL only works with HDFS

  : Your have team of data scientist and you are facing lot of problems, while working with high volume of data. Hence, one of the solution architect in your company
1. A,B,C
2. B,C,D
3. Access Mostly Uused Products by 50000+ Subscribers
4. A,B,C,D

Correct Answer : Get Lastest Questions and Answer :
Explanation: Big SQL, IBM's SQL interface to its Hadoop-based platform, InfoSphere BigInsights. Big SQL is designed to provide SQL developers with an easy on-ramp for
querying data managed by Hadoop. It enables data administrators to create new tables for data stored in Hive, HBase, or their BigInsights distributed file system. In addition, a
LOAD command enables administrators to populate Big SQL tables with data from various sources. And Big SQL's JDBC and ODBC drivers enable many existing tools to use Big SQL to
query this distributed data.
Yet Big SQL doesn't turn Hadoop into one big distributed relational database. If you're curious about what Big SQL can do, we'll explore the basics here, try to clear up some
common misconceptions, and answer many of the questions that relational DBMS users often have about this new technology.

The SQL query engine supports joins, unions, grouping, common table expressions, windowing functions, and other familiar SQL expressions. Furthermore, you can influence the data
access strategy for your queries with optimization hints and configuration options. Depending on the nature of your query, your data volumes, and other factors, Big SQL can use
Hadoop's MapReduce framework to process various query tasks in parallel or execute your query locally within the Big SQL server on a single node whichever may be most
appropriate for your query.





Question : You are looking for a solution, through which you can get deeper, more meaningful insights from your data and predict what is likely to happen next? Which of the
following solution from IBM will help you to get solve this problem
  : You are looking for a solution, through which you can get deeper, more meaningful insights from your data and predict what is likely to happen next? Which of the
1. BigInsight

2. InfoStream

3. Access Mostly Uused Products by 50000+ Subscribers

4. IBM SPSS

Correct Answer : Get Lastest Questions and Answer :
Explanation: What if you could get deeper, more meaningful insights from your data and predict what is likely to happen next? IBM SPSS predictive analytics software
offers advanced techniques in an easy-to-use package to help you find new opportunities, improve efficiency and minimize risk.



Related Questions


Question : Suppose cluster resources become scarce and the scheduler ..
 : Suppose cluster resources become scarce and the scheduler ..
1. Stop working
2. All jobs will hangs for some time
3. All jobs will be stopped and needs to be re-started
4. ResourceManager symmetrically request back resources from a running application


Question :

The ____________ supports a number of features such as weights on queues (heavier queues get more containers), minimum shares, maximum shares, and FIFO policy within queues, but
the basic idea is to share the resources as uniformly as possible.



 :
1. Fair Scheduler
2. Capacity Scheduler
3. FIFO Scheduler
4. Both 1 and 2
5. Both 2 and 3


Question : In your cluster ResourceManager is configured with the "Fair Scheduler", and on average every hour Hadoop runs jobs in parallel.
Now currently single job is running, how much of the resource capacity of the cluster will be used by this running single job.


 : In your cluster ResourceManager is configured with the
1. 1/100 resource of the cluster
2. 20% of the cluster capacity
3. May use full capacity of the cluster
4. It can not be found




Question : Which of the following scheduler ensures that certain users, groups, or production applications always get sufficient resources.
When a queue contains waiting applications, it gets at least its minimum share
 : Which of the following scheduler ensures that certain users, groups, or production applications always get sufficient resources.
1. Fair Scheduler
2. Capacity Scheduler
3. FIFO Scheduler
4. Both 1 and 2
5. Both 2 and 3



Question :

Select the correct statement for the Fair Scheduler



 :
1. allows containers to request variable amounts of memory and schedules based on those requirements
2. If an application is given a container that it cannot use immediately due to a shortage of memory, it can reserve that container, and no other application can use it until the container is released.
3. works best when the workloads are well known, which helps in assigning the minimum capacity.
4. 1 and 2
5. 1 and 3


Question : The Fair scheduler works best when there is a

 : The Fair scheduler works best when there is a
1. When there is a need of Higher Memory
2. lot of variability between queues
3. workloads are well known
4. When there is a need of Higher CPU
5. When all the Jobs needs to be processed in submission order