Apache Hadoop offers a scalable, flexible and reliable distributed computing big data framework for a cluster of systems with storage capacity and local computing power by leveraging commodity hardware. Hadoop follows a Master Slave architecture for the transformation and analysis of large datasets using Hadoop MapReduce paradigm. The 3 important hadoop components that play a vital role in the Hadoop architecture are -
Hadoop Distributed File System (HDFS) – Patterned after the UNIX file system Hadoop MapReduce Yet Another Resource Negotiator (YARN)