Correct Answer : Get Lastest Questions and Answer : Explanation: InputFormat : Hadoop relies on the input format of the job to do three things: 1. Validate the input configuration for the job (i.e., checking that the data is there). 2. Split the input blocks and files into logical chunks of type InputSplit, each of which is assigned to a map task for processing. 3. Access Mostly Uused Products by 50000+ Subscribers
Question : Select the correct statement regarding input split and block size
Correct Answer : Get Lastest Questions and Answer : Explanation: Block is the physical representation of data. Split is the logical representation of data present in Block.
Block and split size can be changed in properties.
Map reads data from Block through splits i.e. split act as a broker between Block and Mapper.
Consider two blocks:
Block 1
aa bb cc dd ee ff gg hh ii jj Block 2
ww ee yy uu oo ii oo pp kk ll nn Now map reads block 1 till aa to JJ and doesn't know how to read block 2 i.e. block doesn't know how to process different block of information. Here comes a Split it will form a Logical grouping of Block 1 and Block 2 as single Block, then it forms offset(key) and line (value) using inputformat and record reader and send map to process further processing.
If your resource is limited and you want to limit the number of maps you can increase the split size. For example: If we have 640 MB of 10 blocks i.e. each block of 64 MB and resource is limited then you can mention Split size as 128 MB then then logical grouping of 128 MB is formed and only 5 maps will be executed with a size of 128 MB.
If we specify split size is false then whole file will form one input split and processed by one map which it takes more time to process when file is big.
Question : Which of the following are the methods available in a InputFormat class and needs to be implemented?
Correct Answer : Get Lastest Questions and Answer : Explanation: InputSplit[] getSplits(JobConf job, int numSplits) throws IOException Logically split the set of input files for the job. Each InputSplit is then assigned to an individual Mapper for processing.
Note: The split is a logical split of the inputs and the input files are not physically split into chunks. For e.g. a split could be tuple.
RecordReader getRecordReader(InputSplit split, JobConf job, Reporter reporter) throws IOException Get the RecordReader for the given InputSplit. It is the responsibility of the RecordReader to respect record boundaries while processing the logical split to present a record-oriented view to the individual task.
1. Map files are stored on the namenode and capture the metadata for all blocks on a particular rack. This is how Hadoop is "rack aware" 2. Map files are the files that show how the data is distributed in the Hadoop cluster. 3. Access Mostly Uused Products by 50000+ Subscribers 4. Map files are sorted sequence files that also have an index. The index allows fast data look up.