What is the disadvantage of using multiple reducers with the default HashPartitioner and distributing your workload across you cluster?
A) You will not be able to compress the intermediate data.
B) You will longer be able to take advantage of a Combiner.
C) By using multiple reducers with the default HashPartitioner, output files may not be in globally sorted order.
D) There are no concerns with this approach. It is always advisable to use multiple reduces.
Correct Answer:
Verified
Q6: In a MapReduce job, you want each
Q7: Which best describes how TextInputFormat processes input
Q8: On a cluster running MapReduce v1 (MRv1),
Q9: Which process describes the lifecycle of a
Q10: The Hadoop framework provides a mechanism for
Q12: To process input key-value pairs, your mapper
Q13: You have the following key-value pairs as
Q14: You need to perform statistical analysis in
Q15: Identify the MapReduce v2 (MRv2 / YARN)
Q16: You have written a Mapper which invokes
Unlock this Answer For Free Now!
View this answer and more for free by performing one of the following actions
Scan the QR code to install the App and get 2 free unlocks
Unlock quizzes for free by uploading documents