This document introduces Hadoop, an open-source framework for distributed storage and processing of large datasets across clusters of computers. It discusses how Hadoop uses HDFS for scalable storage and MapReduce for distributed processing. Key components are introduced, including how HDFS stores data in replicated blocks and how MapReduce executes jobs by splitting data, mapping tasks, shuffling, and reducing results. A word count example demonstrates the MapReduce process.