Premium

IBM Certified Data Architect - Big Data Certification Questions and Answers (Dumps and Practice Questions)



Question : You have been hired as a data analytics engineer, by a stock exchange, where more than a TB data generated on daily basis as a trade info. However, during the
trading time limit, you have analyze all the trade within a 5 minute window. It means every 5 minutes , for the data is keep floowing, you need to check whether their is an fraud
trading happening. Which of the following will help you to analyze continuous flow of data in every 5 mins
  : You have been hired as a data analytics engineer, by a stock exchange, where more than a  TB data generated on daily basis as a trade info. However, during the
1. BigInsight Pig Jobs

2. Infosphere Streams should be used

3. Access Mostly Uused Products by 50000+ Subscribers

4. Hadoop based datastorage engine

Correct Answer : Get Lastest Questions and Answer :
Explanation: InfoSphere Streams is designed to uncover meaningful patterns from information in motion (data flows) during a window of minutes to hours. The platform
provides business value by supporting low-latency insight and better outcomes for time-sensitive applications, such as fraud detection or network management. InfoSphere Streams
also can fuse streams, enabling you to derive new insights from multiple streams.





Question : Which of the following is true about InfoSphere Streams?

A. Respond quickly to events and changing business conditions and requirements.
B. Support continuous analysis of data at rates that are orders of magnitude greater than existing systems.
C. Adapt rapidly to changing data forms and types.
D. Manage high availability, heterogeneity, and distribution for the new stream paradigm.
  : Which of the following is true about InfoSphere Streams?
1. A,B
2. B,C
3. Access Mostly Uused Products by 50000+ Subscribers
4. A,C,D
5. A,B,C,D

Correct Answer : Get Lastest Questions and Answer :
Explanation: InfoSphere Streams provides a programming model and IDE for defining data sources, and software analytic modules called operators fused into processing
execution units. It also provides infrastructure to support the composition of scalable stream-processing applications from these components. The main platform components are:

Runtime environment" This includes platform services and a scheduler for deploying and monitoring Streams applications across a single host or set of integrated hosts.

Programming model" You can write Streams applications using the Streams Processing Language (SPL), a declarative language. You use the language to state what you want, and the
runtime environment accepts the responsibility for determining how best to service the request. In this model, a Streams application is represented as a graph that consists of
operators and the streams that connect them.

Monitoring tools and administrative interfaces" Streams applications process data at speeds much higher than those that the normal collection of operating system monitoring
utilities can efficiently handle. InfoSphere Streams provides the tools that can deal with this environment.





Question : A good use of the BigInsights is as a query-ready archival system for your data warehouse to quickly accessdata
  : A good use of the BigInsights is as a query-ready archival system for your data warehouse to quickly accessdata
1. True
2. False

Correct Answer : Get Lastest Questions and Answer :
Explanation: Data warehouses are growing at unprecedented rates that can cause concerns with regards to performance, management, and costs. Not all data in the
warehouse is accessed frequently, but must be accessible when needed. This data is often archived in the warehouse, which makes it less accessible and more difficult to interact
with.
You can use BigInsights as a query-ready archival system for your data warehouse to quickly access data that is typically archived. Data can be offloaded to BigInsights, where it
can be queried at any time. This implementation saves time and resources by integrating into your existing architecture.



Related Questions


Question : You want to create a BigData Solution, using open source product. Which is having following requirement.

- Text Search Solution, in existing data
- Infrastructure Monitoring is required

Which of the following components can be used

A. HBase
B. Lucene
C. Nagios
D. OOzie
E. Spark
  : You want to create a BigData Solution, using open source product. Which is having following requirement.
1. A,B
2. B,C
3. Access Mostly Uused Products by 50000+ Subscribers
4. D,E
5. A,E


Question : A large global enterprise customer has a Big Data environment set up on Hadoop.
After a year in operation they are now looking to extend access to multiple
functions that will need different views into different aspects/portions of the data.
As you consider these requirements, which of the following statements is TRUE
and also applies to the scenario?
  : A large global enterprise customer has a Big Data environment set up on Hadoop.
1. Hadoop does not support multi tenancy but can easily scale to support this by replicating data to new clusters with commodity hardware.

2. Hadoop can support multi tenancy but only if YARN is used, so if not already used, the customer will need to upgrade to a YARN supported version.

3. Access Mostly Uused Products by 50000+ Subscribers

4. Hadoop can support multi tenancy by using a distributed file system for storage, allowing all nodes to access the data.



Question : What term applies to the data elements in Infosphere Streams?
  : What term applies to the data elements in Infosphere Streams?
1. Tuples

2. Operators

3. Access Mostly Uused Products by 50000+ Subscribers

4. Composite operators


Question : The NameNode uses a file in its _______ to store the EditLog.
  : The NameNode uses a file in its _______  to store the EditLog.
1. Any HDFS Block
2. metastore
3. Access Mostly Uused Products by 50000+ Subscribers
4. local hdfs block



Question : Select the correct option
  : Select the correct option
1. When a file is deleted by a user or an application, it is immediately removed from HDFS
2. When a file is deleted by a user or an application, it is not immediately removed from HDFS. Instead, HDFS first renames it to a file in the /trash directory.
3. Access Mostly Uused Products by 50000+ Subscribers
4. 1,2
5. 2,3


Question : You have data already stored in HDFS and are considering using HBase. Which additional feature does HBase provide to HDFS?

  : You have data already stored in HDFS and are considering using HBase. Which additional feature does HBase provide to HDFS?
1. Random writes
2. Fault tolerance
3. Access Mostly Uused Products by 50000+ Subscribers
4. Batch processing
5. 2,3