Professional Documents
Culture Documents
Hadoop Installation On AWS
Hadoop Installation On AWS
172.31.81.15 ec2-52-203-39-53.compute-1.amazonaws.com
172.31.85.92 ec2-54-89-210-61.compute-1.amazonaws.com
172.31.84.111 ec2-34-230-71-79.compute-1.amazonaws.com
172.31.83.249 ec2-34-201-244-234.compute-1.amazonaws.com
--------------------------------------------------------------------
Host namenode
HostName ec2-52-203-39-53.compute-1.amazonaws.com
User ubuntu
IdentityFile ~/.ssh/mkey.pem
Host datanode1
HostName ec2-54-89-210-61.compute-1.amazonaws.com
User ubuntu
IdentityFile ~/.ssh/mkey.pem
Host datanode2
HostName ec2-34-230-71-79.compute-1.amazonaws.com
User ubuntu
IdentityFile ~/.ssh/mkey.pem
Host datanode3
HostName ec2-34-201-244-234.compute-1.amazonaws.com
User ubuntu
IdentityFile ~/.ssh/mkey.pem
------------------------------------------------------
# namenode
--------------------------------
# Copy the generated keys from id_rsa.pub to authorized_keys in .ssh folder on datanode1,2
------------------------------------------
# All Nodes
java -version
------------------------------------------
wget http://www-us.apache.org/dist/hadoop/common/hadoop-2.7.5/hadoop-2.7.5.tar.gz
sudo tar xzf hadoop-2.7.5.tar.gz
ls
hadoop-2.7.5 hadoop-2.7.5.tar.gz
-----------------------------------------------
-----------------------------------------------
# export JAVA_HOME=/ur
export JAVA_HOME=/usr/lib/jvm/java-1.8.0-openjdk-amd64
export PATH=$PATH:$JAVA_HOME/bin
export HADOOP_HOME=/usr/local/hadoop
export PATH=$PATH:$HADOOP_HOME/bin
export HADOOP_CONF_DIR=/usr/local/hadoop/etc/hadoop
# Java Variable
export HADOOP_CLASSPATH=${JAVA_HOME}/lib/tools.jar
# Hive Variables
export HIVE_HOME=/usr/local/hive
export PATH=$PATH:$HIVE_HOME/bin
export CLASSPATH=$CLASSPATH:/usr/local/hadoop/lib/*:.
export CLASSPATH=$CLASSPATH:/usr/local/hive/lib/*:.
# Load variables
. ~/.profile
# Verify paths
echo $JAVA_HOME
echo $HADOOP_CONF_DIR
echo $HADOOP_HOME
-------------------------------------------
------------------------------
export JAVA_HOME=/usr/lib/jvm/java-1.8.0-openjdk-amd64
----------------------------------
## HADOOP_CONF_DIR/core-site.xmls
<property>
<name>fs.defaultFS</name>
<value>hdfs://ec2-52-203-39-53.compute-1.amazonaws.com:9000</value>
</property>
-------------------------------------------------------------
# Changing /etc/hosts
# PrivateIP PublicDNS
172.31.81.15 ec2-52-203-39-53.compute-1.amazonaws.com
172.31.85.92 ec2-54-89-210-61.compute-1.amazonaws.com
172.31.84.111 ec2-34-230-71-79.compute-1.amazonaws.com
172.31.83.249 ec2-34-201-244-234.compute-1.amazonaws.com
# return ownership
--------------------------------------------------
<configuration>
<property>
<name>dfs.replication</name>
<value>3</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>file:///usr/local/hadoop/hadoop_data/hdfs/namenode</value>
</property>
</configuration>
-----------------------------------------------
# Verify
sudo rm $HADOOP_CONF_DIR/slaves
# Verify
--------------------------------------------------------------
----------------------------------------------------
# on each datanode
------------------------------------------------------
<configuration>
<property>
<name>dfs.replication</name>
<value>3</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>file:///usr/local/hadoop/hadoop_data/hdfs/datanode</value>
</property>
</configuration>
--------------------------------
--------------------------------------
----------------------------------
ec2-52-203-39-53.compute-1.amazonaws.com:50070
OR
namenode_public_ip:50070
-----------------------------------------
$HADOOP_HOME/sbin/start-yarn.sh
-----------------------------------
----------------------------------
$HADOOP_HOME/sbin/stop-yarn.sh
$HADOOP_HOME/sbin/stop-dfs.sh
----------------------------------------
# if it not empty
# generic command