Select the correct option for the flume architecture
1. The Master holds configuration information for each node 2. Node communicate with the Master in every 5 seconds 3. Nodes passes its version number to master 4. 1 and 3 are correct 5. 1,2 and 3
Correct Answer : 5
Question :
Which framework is used for export and import the data between HDFS and RDBMS 1. Flume 2. Sqoop 3. Hive 4. Pig
Correct Answer : 2
Sqoop provides a method to import data from tables in a relational database into HDFS - Does this very efficiently via a Map only MapReduce job - Can also go the other way - Populate database tables from files in HDFS
Refer HadoopExam.com Recorded Training Module : 19
Question : To Create a MapReduce Jobs workflow which framework can be used ?
1. Sqoop 2. Hive 3. Pig 4. OOzie
Correct Answer : 4
Apache OOzie :
Oozie is a workflow engine - Runs on a server - Typically outside the cluster - Runs workflows of Hadoop jobs - Including Pig, Hive, Sqoop jobs - Submits those jobs to the cluster based on a workflow definition - Workflow definitions are submited via HTTP - Jobs can be run at specific times - One time or recurring jobs - Jobs can be run when data is present in a directory