In this tutorial, we will show you how to install Apache Hadoop on Ubuntu 14.04. For those of you who didn’t know, Apache Hadoop is an open-source software framework written in Java for distributed storage and distribution processes, it handles very large size of data sets by distributing them across computer clusters. Rather than rely on hardware to deliver high availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster of computers, each of which may be prone to failures.
This article assumes you have at least basic knowledge of Linux, know how to use the shell, and most importantly, you host your site on your own VPS. The installation is quite simple and assumes you are running in the root account, if not you may need to add ‘
sudo‘ to the commands to get root privileges. I will show you through the step-by-step installation of Apache Hadoop on Ubuntu 14.04. You can follow the same instructions for any other Debian-based distribution like Linux Mint.
Install Apache Hadoop on Ubuntu 14.04
Step 1. Install Java (OpenJDK).
Since Hadoop is based on java, make sure you have Java JDK installed on the system. If you don’t have Java installed on your system, use the following link to install it first.
[email protected] ~# java -version java version "1.8.0_45" Java(TM) SE Runtime Environment (build 1.8.0_45-b14) Java HotSpot(TM) 64-Bit Server VM (build 25.45-b02, mixed mode)
Step 2. Disabling IPv6.
As of now Hadoop does not support IPv6 and is tested to work only on IPv4 networks. If you are using IPv6, you need to switch Hadoop host machines to use IPv4:
Add these 3 lines at the end of the file:
#disable ipv6; net.ipv6.conf.all.disable_ipv6 = 1 net.ipv6.conf.default.disable_ipv6 = 1 net.ipv6.conf.lo.disable_ipv6 = 1
Step 3. Install Apache Hadoop.
To avoid security issues, we recommend setting up a new Hadoop user group and user account to deal with all Hadoop related activities, following command:
sudo addgroup hadoopgroup sudo adduser —ingroup hadoopgroup hadoopuser
After creating a user, it also required to set up key-based ssh to its own account. To do this use execute the following commands:
su - hadoopuser ssh-keygen -t rsa -P "" cat /home/hadoopuser/.ssh/id_rsa.pub >> /home/hadoopuser/.ssh/authorized_keys chmod 600 authorized_keys ssh-copy-id -i ~/.ssh/id_rsa.pub slave-1 ssh slave-1
Download the latest stable version of Apache Hadoop, At the moment of writing this article it is version 2.7.0:
wget http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.7.0/hadoop-2.7.0.tar.gz tar xzf hadoop-2.7.0.tar.gz mv hadoop-2.7.0 hadoop
Step 4. Configure Apache Hadoop.
Setup Hadoop environment variables. Edit
~/.bashrc file and append the following values at end of the file:
export HADOOP_HOME=/home/hadoop/hadoop export HADOOP_INSTALL=$HADOOP_HOME export HADOOP_MAPRED_HOME=$HADOOP_HOME export HADOOP_COMMON_HOME=$HADOOP_HOME export HADOOP_HDFS_HOME=$HADOOP_HOME export YARN_HOME=$HADOOP_HOME export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native export PATH=$PATH:$HADOOP_HOME/sbin:$HADOOP_HOME/bin
Apply environmental variables to the currently running session:
$HADOOP_HOME/etc/hadoop/hadoop-env.sh file and set JAVA_HOME environment variable:
Hadoop has many configuration files, which need to configure as per the requirements of your Hadoop infrastructure. Let’s start with the configuration with basic Hadoop single node cluster setup:
<configuration> <property> <name>fs.default.name</name> <value>hdfs://localhost:9000</value> </property> </configuration>
<configuration> <property> <name>dfs.replication</name> <value>1</value> </property> <property> <name>dfs.name.dir</name> <value>file:///home/hadoop/hadoopdata/hdfs/namenode</value> </property> <property> <name>dfs.data.dir</name> <value>file:///home/hadoop/hadoopdata/hdfs/datanode</value> </property> </configuration>
<configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> </configuration>
<configuration> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> </configuration>
Now format namenode using the following command, do not forget to check the storage directory:
hdfs namenode -format
Start all Hadoop services use the following command:
cd $HADOOP_HOME/sbin/ start-dfs.sh start-yarn.sh
You should observe the output to ascertain that it tries to start the data node on slave nodes one by one. To check if all services are started well use ‘
Step 5. Accessing Apache Hadoop.
Apache Hadoop will be available on HTTP port 8088 and port 50070 by default. Open your favorite browser and navigate to
http://server-ip:50070. If you are using a firewall, please open ports 8088 and 50070 to enable access to the control panel.
Browse the web interface for the ResourceManager by default it is available at
Congratulations! You have successfully installed Apache Hadoop. Thanks for using this tutorial for installing Apache Hadoop on Ubuntu 14.04 system. For additional help or useful information, we recommend you check the official Apache Hadoop website.