This document provides an introduction to MapReduce and the Hadoop Distributed File System (HDFS). It discusses how MapReduce was developed as a programming model for processing large datasets in a distributed manner across clusters. HDFS is introduced as a supporting file system that allows storing large datasets across commodity servers and is optimized for MapReduce workloads. Examples are provided of how MapReduce programs work by defining mapping and reducing functions.