Changes

Jump to: navigation, search

GPU621/Apache Spark

81 bytes added, 16:55, 30 November 2020
Hadoop MapReduce
=== Hadoop MapReduce ===
The processing component of Hadoop ecosystem. It assigns the data fragments from the HDFS to separate map tasks in the cluster and processes the chunks in parallel to combine the pieces into the desired result.
[[File: MapReduce.PNG|thumb|upright=1.2|right|alt=Hadoop cluster|3.2 MapReduce]]
== Applications ==

Navigation menu