This document provides an introduction and overview of Apache Hadoop. It begins with an outline and discusses why Hadoop is important given the growth of data. It then describes the core components of Hadoop - HDFS for distributed storage and MapReduce for distributed computing. The document explains how Hadoop is able to provide scalability and fault tolerance. It provides examples of how Hadoop is used in production at large companies. It concludes by discussing the Hadoop ecosystem and encouraging questions.