In today's data-driven world, organisations generate vast amounts of data daily. Managing and processing this massive volume of structured and unstructured data efficiently requires a robust framework. Apache Hadoop, an open-source distributed storage and processing system, has emerged as a powerful solution for handling big data. Its architecture is built on three key components: HDFS (Hadoop Distributed File System), MapReduce, and YARN (Yet Another Resource Negotiator). This blog explores these fundamental elements and their roles in Hadoop’s ecosystem.