Quantcast
Channel: admin – Robert Chen
Viewing all articles
Browse latest Browse all 18

Multinodes Hadoop 2.4 installation and configuration

$
0
0

 

Prepare the environment

Setup the Hadoop user on all nodes:

root@ubuntu:~# useradd -d /home/hadoop -s /bin/bash -u 600 hadoop

 

add these to /etc/hosts on all 3 hosts:

192.168.47.139 nn

192.168.47.140 dn1

192.168.47.141 dn2

 

unzip JDK, and copy to /usr/local/jdk1.7.0_55

 

vi /etc/profile, add

JAVA_HOME=/usr/local/jdk1.7.0_55

PATH=$PATH:$JAVA_HOME/bin

CLASSPATH=$JAVA_HOME/lib:$JAVA_HOME/jre/lib

export JAVA_HOME CLASSPATHPATH

 

copy ssh keys to data nodes

hadoop@nn:~/.ssh$ ssh-copy-id -i id_rsa.pub hadoop@dn2

/usr/bin/ssh-copy-id: INFO: attempting to log in with the new key(s), to filter out any that are already installed

/usr/bin/ssh-copy-id: INFO: 1 key(s) remain to be installed — if you are prompted now it is to install the new keys

hadoop@dn2’s password:

 

Number of key(s) added: 1

 

Now try logging into the machine, with: “ssh ‘hadoop@dn2′”

and check to make sure that only the key(s) you wanted were added.

 

hadoop@nn:~/.ssh$ ssh-copy-id -i id_rsa.pub hadoop@dn1

/usr/bin/ssh-copy-id: INFO: attempting to log in with the new key(s), to filter out any that are already installed

/usr/bin/ssh-copy-id: INFO: 1 key(s) remain to be installed — if you are prompted now it is to install the new keys

hadoop@dn1’s password:

 

Number of key(s) added: 1

 

Install Hadoop

Download and unzip on all nodes

hadoop@nn:~$ wget http://apache.mirror.iweb.ca/hadoop/common/hadoop-2.4.0/hadoop-2.4.0.tar.gz

 

hadoop@nn:~$ scp hadoop-2.4.0.tar.gz dn1:/home/hadoop/

hadoop-2.4.0.tar.gz 100% 15MB 14.6MB/s 00:00

hadoop@nn:~$ scp hadoop-2.4.0.tar.gz dn2:/home/hadoop/

hadoop-2.4.0.tar.gz 100% 15MB 14.6MB/s 00:00

 

change JAVA_HOME in Hadoop-env.sh

hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ vi hadoop-env.sh

#export JAVA_HOME=${JAVA_HOME}

export JAVA_HOME=/usr/local/jdk1.7.0_55

 

hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ scp hadoop-env.sh dn1:/home/hadoop/hadoop-2.4.0/etc/hadoop/

hadoop-env.sh 100% 3535 3.5KB/s 00:00

hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ scp hadoop-env.sh dn2:/home/hadoop/hadoop-2.4.0/etc/hadoop/

hadoop-env.sh 100% 3535 3.5KB/s 00:00

 

update the .bash_profile of Hadoop user

hadoop@nn:~$ more .bash_profile

export HADOOP_HOME=/home/hadoop/hadoop-2.4.0

export PATH=/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games:/usr/local/jdk1.7.0_55/bin:/home/hadoop/hadoop-2.4

.0/bin:/home/hadoop/hadoop-2.4.0/sbin

alias h=’cd /home/hadoop/hadoop-2.4.0/’

alias etc=’cd /home/hadoop/hadoop-2.4.0//etc/hadoop’

export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native

export HADOOP_OPTS=”-Djava.library.path=$HADOOP_HOME/lib/native”

export CLASSPATH=$CLASSPATH:$HADOOP_HOME/share/hadoop/common:$HADOOP_HOME/share/hadoop/common/lib:\

$HADOOP_HOME/share/hadoop/yarn:$HADOOP_HOME/share/hadoop/yarn/lib:\

$HADOOP_HOME/share/hadoop/mapreduce:$HADOOP_HOME/share/hadoop/mapreduce/lib:\

$HADOOP_HOME/share/hadoop/hdfs:$HADOOP_HOME/share/hadoop/hdfs/lib

 

Write configurations

Core-site.xml

<configuration>

<property>

<name>fs.default.name</name>

<value>hdfs://nn:9000</value>

</property>

<property>

<name>hadoop.tmp.dir</name>

<value>/home/hadoop/hdfs/tmp</value>

</property>

<property>

<name>fs.checkpoint.dir</name>

<value>/home/hadoop/hdfs/checkpoint</value>

</property>

<property>

<name>hadoop.logfile.size</name>

<value>104857600</value>

</property>

<property>

<name>hadoop.logfile.count</name>

<value>20</value>

</property>

<property>

<name>io.file.buffer.size</name>

<value>131072</value>

</property>

</configuration>

 

Hdfs-site.xml

 

<configuration>

<property>

<name>dfs.name.dir</name>

<value>/home/hadoop/hdfs/name</value>

<description>Path on the local filesystem where the NameNode stores the namespace and transactions logs persistently.</description>

</property>

<property>

<name>dfs.data.dir</name>

<value>/home/hadoop/hdfs/data</value>

<description>Comma separated list of paths on the local filesystem of a DataNode where it should store its blocks.</description>

</property>

<property>

<name>dfs.replication</name>

<value>2</value>

</property>

<property>

<name>dfs.datanode.du.reserved</name>

<value>0</value>

</property>

</configuration>

 

Mapred-site.xml

 

<configuration>

<property>

<name>mapreduce.framework.name</name>

<value>yarn</value>

</property>

<property>

<name>mapreduce.map.memory.mb</name>

<value>1536</value>

</property>

<property>

<name>mapred.child.java.opts</name>

<value>-Xmx4096m</value>

</property>

<property>

<name>mapreduce.map.java.opts</name>

<value>-Xmx1228m</value>

<property>

</property>

<name>mapreduce.reduce.memory.mb</name>

<value>3072</value>

</property>

<property>

<name>mapreduce.reduce.java.opts</name>

<value>-Xmx2457m</value>

</property>

<property>

<name>mapreduce.reduce.shuffle.parallelcopies</name>

<value>50</value>

</property>

<property>

<name>mapreduce.task.io.sort.mb</name>

<value>614</value>

</property>

<property>

<name>mapreduce.task.io.sort.factor</name>

<value>100</value>

</property>

<property>

<name>mapreduce.jobhistory.address</name>

<value>nn:10020</value>

</property>

<property>

<name>mapreduce.jobhistory.webapp.address</name>

<value>nn:19888</value>

</property>

</configuration>

 

Yarn-site.xml

<configuration>

<property>

<name>yarn.resourcemanager.address</name>

<value>nn:8132</value>

</property>

<property>

<name>yarn.resourcemanager.scheduler.address</name>

<value>nn:8130</value>

</property>

<property>

<name>yarn.resourcemanager.resource-tracker.address</name>

<value>nn:8131</value>

</property>

<property>

<name>yarn.resourcemanager.webapp.address</name>

<value>nn:8188</value>

</property>

<property>

<name>yarn.resourcemanager.amin.address</name>

<value>nn:8288</value>

</property>

<property>

<name>yarn.scheduler.minimum-allocation-mb</name>

<value>1536</value>

</property>

<property>

<name>yarn.scheduler.maximum-allocation-mb</name>

<value>43008</value>

</property>

<property>

<name>yarn.nodemanager.resource.memory-mb</name>

<value>43008</value>

</property>

<property>

<name>yarn.scheduler.minimum-allocation-vcores</name>

<value>1</value>

</property>

<property>

<name>yarn.scheduler.maximum-allocation-vcores</name>

<value>12</value>

</property>

<property>

<name>yarn.app.mapreduce.am.resource.mb</name>

<value>3072</value>

</property>

<property>

<name>yarn.app.mapreduce.am.command-opts</name>

<value>-Xmx2457m</value>

</property>

<property>

<name>yarn.resourcemanager.scheduler.class</name>

<value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>

</property>

<property>

<name>yarn.nodemanager.aux-services</name>

<value>mapreduce_shuffle</value>

</property>

<property>

<name>yarn.nodemanager.local-dirs</name>

<value>${hadoop.tmp.dir}/nodemanager/local</value>

</property>

<property>

<name>yarn.nodemanager.address</name>

<value>0.0.0.0:8034</value>

<description>the nodemanagers bind to this port</description>

</property>

<property>

<name>yarn.nodemanager.remote-app-log-dir</name>

<value>${hadoop.tmp.dir}/nodemanager/remote</value>

<description>directory on hdfs where the application logs are moved to </description>

</property>

<property>

<name>yarn.app.mapreduce.am.command-opts</name>

<value>-Xmx2457m</value>

</property>

<property>

<name>yarn.resourcemanager.scheduler.class</name>

<value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>

</property>

<property>

<name>yarn.nodemanager.aux-services</name>

<value>mapreduce_shuffle</value>

</property>

<property>

<name>yarn.nodemanager.local-dirs</name>

<value>${hadoop.tmp.dir}/nodemanager/local</value>

</property>

<property>

<name>yarn.nodemanager.address</name>

<value>0.0.0.0:8034</value>

<description>the nodemanagers bind to this port</description>

</property>

<property>

<name>yarn.nodemanager.remote-app-log-dir</name>

<value>${hadoop.tmp.dir}/nodemanager/remote</value>

<description>directory on hdfs where the application logs are moved to </description>

</property>

<property>

<name>yarn.nodemanager.log-dirs</name>

<value>${hadoop.tmp.dir}/nodemanager/logs</value>

<description>the directories used by Nodemanagers as log directories</description>

</property>

<property>

<name>yarn.application.classpath</name>

<value> /home/hadoop/hadoop-2.4.0/etc/hadoop,/home/hadoop/hadoop-2.4.0/share/hadoop/common/*,/home/hadoop/hadoop-2.4.0/share/hadoop/common/lib/*,/home/hadoop/hadoop-2.4.0/share/hadoop/hdfs/*,/home/hadoop/hadoop-2.4.0/share/hadoop/hdfs/lib/*,/home/hadoop/hadoop-2.4.0/share/hadoop/mapreduce/*,/home/hadoop/hadoop-2.4.0/share/hadoop/mapreduce/lib/*,/home/hadoop/hadoop-2.4.0/share/hadoop/yarn/*,/home/hadoop/hadoop-2.4.0/share/hadoop/yarn/lib/*

</value>

</property>

<property>

<name>yarn.log-aggregation-enable</name>

<value>true</value>

</property>

</configuration>

 

Setup slaves, master

hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ more slaves

dn1

dn2

hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ more master

nn

 

copy files to other nodes

copy slaves, master to dn1, dn2

 

hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ scp mapred-site.xml core-site.xml hdfs-site.xml yarn-site.xml dn1:/home/hadoop/hadoop-2.4.0/etc/hadoop/

mapred-site.xml 100% 2028 2.0KB/s 00:00

core-site.xml 100% 1420 1.4KB/s 00:00

hdfs-site.xml 100% 1590 1.6KB/s 00:00

yarn-site.xml 100% 5301 5.2KB/s 00:00

hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ scp mapred-site.xml core-site.xml hdfs-site.xml yarn-site.xml dn2:/home/hadoop/hadoop-2.4.0/etc/hadoop/

mapred-site.xml 100% 2028 2.0KB/s 00:00

core-site.xml 100% 1420 1.4KB/s 00:00

hdfs-site.xml 100% 1590 1.6KB/s 00:00

yarn-site.xml 100% 5301 5.2KB/s 00:00

 

on the master, start namenode, resourcemanager and etc

136 hadoop namenode -format

137 hadoop-daemon.sh –script hdfs start namenode

138 ps -ef|grep hadoop

139 yarn-daemon.sh start resourcemanager

140 ps -ef|grep hadoop

141 yarn-daemon.sh start proxyserver

142 mr-jobhistory-daemon.sh start historyserver

 

On the data nodes, starts datanode, nodemanager

hadoop@dn1:~/hadoop-2.4.0/logs$ hadoop-daemon.sh –script hdfs start datanode

starting datanode, logging to /home/hadoop/hadoop-2.4.0/logs/hadoop-hadoop-datanode-dn1.out

hadoop@dn1:~/hadoop-2.4.0/logs$ yarn-daemon.sh start nodemanager

starting nodemanager, logging to /home/hadoop/hadoop-2.4.0/logs/yarn-hadoop-nodemanager-dn1.out

hadoop@dn2:~$ hadoop-daemon.sh –script hdfs start datanode

starting datanode, logging to /home/hadoop/hadoop-2.4.0/logs/hadoop-hadoop-datanode-dn2.out

hadoop@dn2:~$ yarn-daemon.sh start nodemanager

starting nodemanager, logging to /home/hadoop/hadoop-2.4.0/logs/yarn-hadoop-nodemanager-dn2.out

 

check status

hadoop@nn:~$ jps

4359 ResourceManager

6691 Jps

4275 NameNode

4652 JobHistoryServer

 

hadoop@dn2:~$ jps

4033 DataNode

4089 NodeManager

5392 Jps

 

Fix the WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform

Recompile with source codes and replace the libs

configure: error: C++ preprocessor “/lib/cpp” fails sanity check

solution: apt-get install g++

 

root@nn:~/Downloads/protobuf-2.5.0# protoc –version

protoc: error while loading shared libraries: libprotoc.so.8: cannot open shared object file: No such file or directory

solution: add this to /etc/profile

export LD_LIBRARY_PATH=/usr/local/lib:/lib64:/usr/lib64

root@nn:~/Downloads/protobuf-2.5.0# source /etc/profile

root@nn:~/Downloads/protobuf-2.5.0# protoc –version

 

apt-get install zlib1g-dev libncurses5-dev

apt-get install gcc build-essential libssl-dev zlib1g-dev libglib2.0-dev cmake maven

hadoop@nn:~/hadoop-2.4.0-src$ mvn clean package -Pdist,native -DskipTests –Dtar


[INFO] BUILD SUCCESS

[INFO] ————————————————————————

[INFO] Total time: 17:18.833s

[INFO] Finished at: Tue May 27 07:55:58 PDT 2014

[INFO] Final Memory: 91M/239M

[INFO] ————————————————————————

 

hadoop@nn:~/hadoop-2.4.0-src/hadoop-common-project/hadoop-common/target/native/target/usr/local/lib$ ls -l

total 1732

-rw-rw-r– 1 hadoop hadoop 1111258 May 27 07:39 libhadoop.a

lrwxrwxrwx 1 hadoop hadoop 18 May 27 07:39 libhadoop.so -> libhadoop.so.1.0.0

-rwxrwxr-x 1 hadoop hadoop 657597 May 27 07:39 libhadoop.so.1.0.0

 

hadoop@nn:~/hadoop-2.4.0-src/hadoop-common-project/hadoop-common/target/native/target/usr/local/lib$ cp * /home/hadoop/hadoop-2.4.0/lib/native/

 

hadoop@nn:~$ cd example

hadoop@nn:~/example$ ls

file1.txt file2.txt

hadoop@nn:~/example$ hadoop fs -mkdir /data

 

hadoop@nn:~/example$ hadoop fs -put -f file1.txt file2.txt /data

map reduce job failed

hadoop@nn:~/hadoop-2.4.0$ hadoop jar ./share/hadoop/mapreduce/sources/hadoop-mapreduce-examples-2.4.0-sources.jar org.apache.hadoop.examples.WordCount /data /output

14/05/27 11:30:54 INFO client.RMProxy: Connecting to ResourceManager at nn/192.168.47.139:8132

14/05/27 11:30:55 INFO input.FileInputFormat: Total input paths to process : 2

14/05/27 11:30:55 INFO mapreduce.JobSubmitter: number of splits:2

14/05/27 11:30:56 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1401214983224_0001

14/05/27 11:30:57 INFO impl.YarnClientImpl: Submitted application application_1401214983224_0001

14/05/27 11:30:57 INFO mapreduce.Job: The url to track the job: http://nn:8188/proxy/application_1401214983224_0001/

14/05/27 11:30:57 INFO mapreduce.Job: Running job: job_1401214983224_0001

14/05/27 11:31:04 INFO mapreduce.Job: Job job_1401214983224_0001 running in uber mode : false

14/05/27 11:31:04 INFO mapreduce.Job: map 0% reduce 0%

14/05/27 11:31:04 INFO mapreduce.Job: Job job_1401214983224_0001 failed with state FAILED due to: Application application_1401214983224_0001 failed 2 times due to AM Container for appattempt_1401214983224_0001_000002 exited with exitCode: 1 due to: Exception from container-launch: org.apache.hadoop.util.Shell$ExitCodeException:

org.apache.hadoop.util.Shell$ExitCodeException:

at org.apache.hadoop.util.Shell.runCommand(Shell.java:505)

at org.apache.hadoop.util.Shell.run(Shell.java:418)

at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)

at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)

at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)

at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)

at java.util.concurrent.FutureTask.run(FutureTask.java:262)

at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)

at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)

at java.lang.Thread.run(Thread.java:745)

 

 

Container exited with a non-zero exit code 1

.Failing this attempt.. Failing the application.

14/05/27 11:31:04 INFO mapreduce.Job: Counters: 0

 

hadoop@nn:~/hadoop-2.4.0$ hadoop jar ./share/hadoop/mapreduce/sources/hadoop-mapreduce-examples-2.4.0-sources.jar org.apache.had oop.examples.WordCount /data /output

14/05/27 11:44:19 INFO client.RMProxy: Connecting to ResourceManager at nn/192.168.47.139:8132

14/05/27 11:44:21 INFO input.FileInputFormat: Total input paths to process : 2

14/05/27 11:44:22 INFO mapreduce.JobSubmitter: number of splits:2

14/05/27 11:44:22 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1401216238271_0001

14/05/27 11:44:22 INFO impl.YarnClientImpl: Submitted application application_1401216238271_0001

14/05/27 11:44:23 INFO mapreduce.Job: The url to track the job: http://nn:8188/proxy/application_1401216238271_0001/

14/05/27 11:44:23 INFO mapreduce.Job: Running job: job_1401216238271_0001

14/05/27 11:44:49 INFO mapreduce.Job: Job job_1401216238271_0001 running in uber mode : false

14/05/27 11:44:49 INFO mapreduce.Job: map 0% reduce 0%

14/05/27 11:44:58 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000000_0, Status : FAILED

Container [pid=7385,containerID=container_1401216238271_0001_01_000002] is running beyond virtual memory limits. Current usage: 67.2 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.

Dump of the process-tree for container_1401216238271_0001_01_000002 :

|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE

|- 7396 7385 7385 7385 (java) 179 18 4726177792 16837 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -D hadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/appli cation_1401216238271_0001/container_1401216238271_0001_01_000002/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app .container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000002 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attemp t_1401216238271_0001_m_000000_0 2

|- 7385 7224 7385 7385 (bash) 0 0 17039360 354 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000002/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000002 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000000_0 2 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000002/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000002/stderr

 

Container killed on request. Exit code is 143

Container exited with a non-zero exit code 143

 

14/05/27 11:45:00 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000001_0, Status : FAILED

Container [pid=7384,containerID=container_1401216238271_0001_01_000003] is running beyond virtual memory limits. Current usage: 68.6 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.

Dump of the process-tree for container_1401216238271_0001_01_000003 :

|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE

|- 7394 7384 7384 7384 (java) 178 18 4726177792 17213 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -D hadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/appli cation_1401216238271_0001/container_1401216238271_0001_01_000003/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app .container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000003 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attemp t_1401216238271_0001_m_000001_0 3

|- 7384 7224 7384 7384 (bash) 0 0 17039360 354 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000003/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000003 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000001_0 3 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000003/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000003/stderr

 

Container killed on request. Exit code is 143

Container exited with a non-zero exit code 143

 

14/05/27 11:45:08 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000000_1, Status : FAILED

Container [pid=7441,containerID=container_1401216238271_0001_01_000004] is running beyond virtual memory limits. Current usage: 59.4 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.

Dump of the process-tree for container_1401216238271_0001_01_000004 :

|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE

|- 7441 7224 7441 7441 (bash) 0 0 17039360 354 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000004/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000004 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000000_1 4 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000004/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000004/stderr

|- 7446 7441 7441 7441 (java) 178 8 4726177792 14843 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -Dh adoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/applic ation_1401216238271_0001/container_1401216238271_0001_01_000004/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app. container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000004 – Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt _1401216238271_0001_m_000000_1 4

 

Container killed on request. Exit code is 143

Container exited with a non-zero exit code 143

 

14/05/27 11:45:08 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000001_1, Status : FAILED

Container [pid=7456,containerID=container_1401216238271_0001_01_000005] is running beyond virtual memory limits. Current usage: 58.4 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.

Dump of the process-tree for container_1401216238271_0001_01_000005 :

|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE

|- 7461 7456 7456 7456 (java) 177 9 4726177792 14603 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -Dh adoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/applic ation_1401216238271_0001/container_1401216238271_0001_01_000005/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app. container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000005 – Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt _1401216238271_0001_m_000001_1 5

|- 7456 7224 7456 7456 (bash) 0 0 17039360 353 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000005/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000005 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000001_1 5 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000005/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000005/stderr

 

Container killed on request. Exit code is 143

Container exited with a non-zero exit code 143

 

14/05/27 11:45:17 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000000_2, Status : FAILED

Container [pid=7089,containerID=container_1401216238271_0001_01_000008] is running beyond virtual memory limits. Current usage: 63.2 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.

Dump of the process-tree for container_1401216238271_0001_01_000008 :

|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE

|- 7089 6803 7089 7089 (bash) 0 1 17039360 353 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000008/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000008 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000000_2 8 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000008/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000008/stderr

|- 7095 7089 7089 7089 (java) 182 25 4726177792 15832 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -D hadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/appli cation_1401216238271_0001/container_1401216238271_0001_01_000008/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app .container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000008 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attemp t_1401216238271_0001_m_000000_2 8

 

Container killed on request. Exit code is 143

Container exited with a non-zero exit code 143

 

14/05/27 11:45:17 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000001_2, Status : FAILED

Container [pid=7107,containerID=container_1401216238271_0001_01_000009] is running beyond virtual memory limits. Current usage: 54.7 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.

Dump of the process-tree for container_1401216238271_0001_01_000009 :

|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE

|- 7113 7107 7107 7107 (java) 180 19 4726177792 13644 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -D hadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/appli cation_1401216238271_0001/container_1401216238271_0001_01_000009/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app .container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000009 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attemp t_1401216238271_0001_m_000001_2 9

|- 7107 6803 7107 7107 (bash) 0 0 17039360 353 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000009/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000009 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000001_2 9 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000009/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000009/stderr

 

Container killed on request. Exit code is 143

Container exited with a non-zero exit code 143

 

14/05/27 11:45:24 INFO mapreduce.Job: map 100% reduce 100%

14/05/27 11:45:24 INFO mapreduce.Job: Job job_1401216238271_0001 failed with state FAILED due to: Task failed task_1401216238271 _0001_m_000001

Job failed as tasks failed. failedMaps:1 failedReduces:0

 

14/05/27 11:45:24 INFO mapreduce.Job: Counters: 13

Job Counters

Failed map tasks=7

Killed map tasks=1

Launched map tasks=8

Other local map tasks=6

Data-local map tasks=2

Total time spent by all maps in occupied slots (ms)=50333

Total time spent by all reduces in occupied slots (ms)=0

Total time spent by all map tasks (ms)=50333

Total vcore-seconds taken by all map tasks=50333

Total megabyte-seconds taken by all map tasks=77311488

Map-Reduce Framework

CPU time spent (ms)=0

Physical memory (bytes) snapshot=0

Virtual memory (bytes) snapshot=0

 

hadoop@nn:~/hadoop-2.4.0$ hadoop jar ./share/hadoop/mapreduce/sources/hadoop-mapreduce-examples-2.4.0-sources.jar org.apache.hadoop.examples.WordCount /data /output4

14/05/27 13:16:57 INFO client.RMProxy: Connecting to ResourceManager at nn/192.168.47.139:8132

14/05/27 13:16:58 INFO input.FileInputFormat: Total input paths to process : 2

14/05/27 13:16:58 INFO mapreduce.JobSubmitter: number of splits:2

14/05/27 13:16:59 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1401221062327_0001

14/05/27 13:16:59 INFO impl.YarnClientImpl: Submitted application application_1401221062327_0001

14/05/27 13:16:59 INFO mapreduce.Job: The url to track the job: http://nn:8188/proxy/application_1401221062327_0001/

14/05/27 13:16:59 INFO mapreduce.Job: Running job: job_1401221062327_0001

14/05/27 13:17:26 INFO mapreduce.Job: Job job_1401221062327_0001 running in uber mode : false

14/05/27 13:17:26 INFO mapreduce.Job: map 0% reduce 0%

14/05/27 13:18:01 INFO mapreduce.Job: map 100% reduce 0%

14/05/27 13:18:30 INFO mapreduce.Job: map 100% reduce 100%

14/05/27 13:18:30 INFO mapreduce.Job: Job job_1401221062327_0001 completed successfully

14/05/27 13:18:30 INFO mapreduce.Job: Counters: 49

File System Counters

FILE: Number of bytes read=103

FILE: Number of bytes written=282013

FILE: Number of read operations=0

FILE: Number of large read operations=0

FILE: Number of write operations=0

HDFS: Number of bytes read=272

HDFS: Number of bytes written=56

HDFS: Number of read operations=9

HDFS: Number of large read operations=0

HDFS: Number of write operations=2

Job Counters

Launched map tasks=2

Launched reduce tasks=1

Data-local map tasks=2

Total time spent by all maps in occupied slots (ms)=65907

Total time spent by all reduces in occupied slots (ms)=25452

Total time spent by all map tasks (ms)=65907

Total time spent by all reduce tasks (ms)=25452

Total vcore-seconds taken by all map tasks=65907

Total vcore-seconds taken by all reduce tasks=25452

Total megabyte-seconds taken by all map tasks=67488768

Total megabyte-seconds taken by all reduce tasks=26062848

Map-Reduce Framework

Map input records=6

Map output records=12

Map output bytes=123

Map output materialized bytes=109

Input split bytes=188

Combine input records=12

Combine output records=8

Reduce input groups=7

Reduce shuffle bytes=109

Reduce input records=8

Reduce output records=7

Spilled Records=16

Shuffled Maps =2

Failed Shuffles=0

Merged Map outputs=2

GC time elapsed (ms)=424

CPU time spent (ms)=3410

Physical memory (bytes) snapshot=486014976

Virtual memory (bytes) snapshot=6566645760

Total committed heap usage (bytes)=256843776

Shuffle Errors

BAD_ID=0

CONNECTION=0

IO_ERROR=0

WRONG_LENGTH=0

WRONG_MAP=0

WRONG_REDUCE=0

File Input Format Counters

Bytes Read=84

File Output Format Counters

Bytes Written=56

 

hadoop@nn:~/hadoop-2.4.0$ hadoop fs -rm -r /output

14/05/27 16:38:41 INFO fs.TrashPolicyDefault: Namenode trash configuration: Deletion interval = 0 minutes, Emptier interval = 0 minutes.

Deleted /output

The post Multinodes Hadoop 2.4 installation and configuration appeared first on Robert Chen.


Viewing all articles
Browse latest Browse all 18

Trending Articles