Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
0/
http://tecadmin.net/setup-hadoop-2-4-single-node-cluster-on-linux/
# java -version
openjdk version "1.8.0_45"
OpenJDK Runtime Environment (build 1.8.0_45-b13)
OpenJDK 64-Bit Server VM (build 25.45-b02, mixed mode)
Install java versi 8
# yum install java-1.8.0-openjdk-devel
Pilih alternatif versi java (optional)
# alternatives --install /usr/bin/java java /usr/java/latest/bin/java 1
# alternatives --config java
Tambahkan dns pada file /etc/hosts dengan nilai <alamat ip> <hostname>,
catatan: alamat ip dan hostname menyesuaikan dari computer masing-masing.
Contoh sebagai berikut:
Alamat ip computer yang digunakan adalah 172.18.107.61
Nama hostname yang digunakan adalah HadoopMstr1
# vim /etc/hosts
127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4
::1
localhost localhost.localdomain localhost6 localhost6.localdomain6
172.18.107.61 HadoopMstr1
# useradd hadoop
# passwd hadoop
# su - hadoop
$ ssh-keygen -t rsa -P '' -f ~/.ssh/id_rsa
$ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
$ chmod 0600 ~/.ssh/authorized_keys
$ ssh localhost
$ exit
$
$
$
$
cd ~
wget http://apache.claz.org/hadoop/common/hadoop-2.6.0/hadoop-2.6.0.tar.gz
tar xzf hadoop-2.6.0.tar.gz
mv hadoop-2.6.0 hadoop
$ ls -l
total 190688
drwxr-xr-x. 9 hadoop hadoop
4096 Nov 14 2014 hadoop
-rw-rw-r--. 1 hadoop hadoop 195257604 May 26 10:02 hadoop-2.6.0.tar.gz
$ vim .bashrc
# .bashrc
# Source global definitions
if [ -f /etc/bashrc ]; then
. /etc/bashrc
fi
<value>yarn</value>
</property>
</configuration>
$ vim $HADOOP_HOME/etc/hadoop/yarn-site.xml
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.class</name>
<value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairSche
duler</value>
<!-<value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoSche
duler</value> -->
<!-<value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.Cap
acityScheduler</value> -->
</property>
<property>
<description>The address of the RM web application.</description>
<name>yarn.resourcemanager.webapp.address</name>
<value>localhost:18088</value>
</property>
<property>
<name>yarn.resourcemanager.resource-tracker.address</name>
<value>localhost:18031</value>
</property>
<property>
<description>The address of the scheduler interface.</description>
<name>yarn.resourcemanager.scheduler.address</name>
<value>localhost:18030</value>
</property>
<property>
<description>The address of the applications manager interface in the
RM.</description>
<name>yarn.resourcemanager.address</name>
<value>localhost:18032</value>
</property>
<property>
<description>The address of the RM admin interface.</description>
<name>yarn.resourcemanager.admin.address</name>
<value>localhost:18033</value>
</property>
<property>
<description>Set to false, to avoid ip check</description>
<name>hadoop.security.token.service.use_ip</name>
<value>false</value>
</property>
<property>
<name>yarn.scheduler.capacity.maximum-applications</name>
<value>1000</value>
<description>Maximum number of applications in the system which
can be concurrently active both running and pending</description>
</property>
<property>
<description>Whether to use preemption. Note that preemption is
experimental
in the current version. Defaults to false.</description>
<name>yarn.scheduler.fair.preemption</name>
<value>true</value>
</property>
<property>
<description>Whether to allow multiple container assignments in one
heartbeat. Defaults to false.</description>
<name>yarn.scheduler.fair.assignmultiple</name>
<value>true</value>
</property>
</configuration
$ vim $HADOOP_HOME/etc/hadoop/yarn-site.xml
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
</configuration
Now format the namenode using following command, make sure that Storage
directory is
$ hdfs namenode -format
15/02/04 09:58:43 INFO namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG: host = java.net.UnknownHostException: HadoopMstr1:
HadoopMstr1
STARTUP_MSG: args = [-format]
STARTUP_MSG: version = 2.6.0
...
...
secondary namenode
http://localhost:50090/
DataNode
http://localhost:50075/
login to root to create directori /var/log/httpd, buat mode akses (chmod 777),
kemudian kembali lagi ke user hadoop. Dan jalankan perintah berikut:
/home/hadoop/hadoop/bin/hdfs dfs -put /var/log/httpd logs
Replay to pete
Hi pete, saya lihat kamu telah berhasil dengan lancar dan tidak ada
permasalahan. saya telah mengikuti semua langkah yang telah dituliskan diatas
tapi saya tidak bisa mengakses DataNode dan cluster and all applications
Hi pete, I see you've managed smoothly and no problems. I have followed all the
steps that have been written above but I can not access DataNode and cluster
and all applications
DataNode:
http://localhost:50075
cluster and all applications:
http://localhost:18088