The Hadoop paradigm originated from Google
and is used in crunching large data sets. It is ideally suited for applications
like Big Data, creating an inverted index used by search engines and other
problems which require terabytes of data to processed in parallel. One key
aspect of Hadoop is that it is made up of commodity servers. Hence server, disk
crashes or network issues are assumed to be norm rather than an exception.
Hadoop paradigm is made of the Map-Reduce & the HDFS parts. The
Map-Reduce has 2 major components to... [More]