This document discusses the growth of data and challenges in storing and analyzing large datasets. It introduces Hadoop as a solution for processing large datasets in parallel across commodity servers. Key aspects of Hadoop covered include its core components HDFS for storage and MapReduce for distributed processing. Example uses by large companies like Amazon and Facebook are listed. The document contrasts Hadoop with RDBMS and explains when Hadoop is preferable to use.