In a MapReduce job, you want each of your input files processed by a single map task. How do you configure a MapReduce job so that a single map task processes each input file regardless of how many blocks the input file occupies?
A) Increase the parameter that controls minimum split size in the job configuration.
B) Write a custom MapRunner that iterates over all key-value pairs in the entire file.
C) Set the number of mappers equal to the number of input files you want to process.
D) Write a custom FileInputFormat and override the method isSplitable to always return false.
Correct Answer:
Verified
Q1: For each intermediate key, each reducer task
Q2: For each input key-value pair, mappers can
Q3: Indentify which best defines a SequenceFile?
A) A
Q4: Determine which best describes when the reduce
Q5: You have user profile records in your
Q7: Which best describes how TextInputFormat processes input
Q8: On a cluster running MapReduce v1 (MRv1),
Q9: Which process describes the lifecycle of a
Q10: The Hadoop framework provides a mechanism for
Q11: What is the disadvantage of using multiple
Unlock this Answer For Free Now!
View this answer and more for free by performing one of the following actions
Scan the QR code to install the App and get 2 free unlocks
Unlock quizzes for free by uploading documents