This tutorial will guide you through the installation process of Apache Hadoop Distributed File System (HDFS) on OpenSUSE Latest.
Before you start, make sure your system meets the following requirements:
First, download the latest stable release of Hadoop from the official website. You can use the following command to download it:
wget https://downloads.apache.org/hadoop/common/hadoop-3.3.1/hadoop-3.3.1.tar.gz
Extract the downloaded Hadoop archive using the following command:
tar -xvf hadoop-3.3.1.tar.gz
Next, navigate to the Hadoop directory and configure it by editing the etc/hadoop/core-site.xml
file:
cd hadoop-3.3.1
nano etc/hadoop/core-site.xml
Add the following configuration properties to the core-site.xml
file:
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://localhost:9000</value>
</property>
</configuration>
Save and exit the file.
Now, start the Hadoop NameNode and DataNode by executing the following command:
sbin/start-dfs.sh
You can verify that the NameNode and DataNode are running by accessing the HDFS status page at http://localhost:50070/
.
Create a directory in HDFS by executing the following command:
bin/hdfs dfs -mkdir /input
Copy some sample data to the HDFS directory by executing the following command:
bin/hdfs dfs -put etc/hadoop/*.xml /input
Verify that the data has been copied to HDFS by executing the following command:
bin/hdfs dfs -ls /input
You should see a list of the XML files that were copied to HDFS.
Congratulations! You have successfully installed HDFS on OpenSUSE Latest and verified its functionality by copying data to it. You can now use HDFS to store and process large amounts of data.
If you want to self-host in an easy, hands free way, need an external IP address, or simply want your data in your own hands, give IPv6.rs a try!
Alternatively, for the best virtual desktop, try Shells!