Hadoop is a Java framework that integrates elements of both the MapReduce computing paradigm and the Google File System (GFS) for executing the application on huge clusters of commodity hardware. Like Hadoop in particular, HDFS is a distributed file system with a high fault tolerance that is made to be installed on inexpensive hardware. It is appropriate for applications with huge data volumes and offers high throughput accessibility to application data.
By default, the Hadoop configuration file may be found in /etc/hadoop/hdfs-site.xml. The dfs. namenode property is located here.
Hadoop 2.x is being widely used and you can learn the same.
Gigabytes to petabytes of data may be stored and processed effectively using the open source framework known as Apache Hadoop. Hadoop enables clustering many computers to examine big datasets in parallelism more swiftly than using a single powerful machine for data storage and processing.
Java-based Hadoop is a framework for running programmes on a sizable cluster of public hardware. It resembles the Google file system in many ways. Java must first be installed in Ubuntu because we require it to install Hadoop.