Image may be NSFW.
Clik here to view.
Image may be NSFW.
Clik here to view.
Prepare the environment
Setup the Hadoop user on all nodes:
root@ubuntu:~# useradd -d /home/hadoop -s /bin/bash -u 600 hadoop
add these to /etc/hosts on all 3 hosts:
192.168.47.139 nn
192.168.47.140 dn1
192.168.47.141 dn2
unzip JDK, and copy to /usr/local/jdk1.7.0_55
vi /etc/profile, add
JAVA_HOME=/usr/local/jdk1.7.0_55
PATH=$PATH:$JAVA_HOME/bin
CLASSPATH=$JAVA_HOME/lib:$JAVA_HOME/jre/lib
export JAVA_HOME CLASSPATHPATH
copy ssh keys to data nodes
hadoop@nn:~/.ssh$ ssh-copy-id -i id_rsa.pub hadoop@dn2
/usr/bin/ssh-copy-id: INFO: attempting to log in with the new key(s), to filter out any that are already installed
/usr/bin/ssh-copy-id: INFO: 1 key(s) remain to be installed — if you are prompted now it is to install the new keys
hadoop@dn2’s password:
Number of key(s) added: 1
Now try logging into the machine, with: “ssh ‘hadoop@dn2′”
and check to make sure that only the key(s) you wanted were added.
hadoop@nn:~/.ssh$ ssh-copy-id -i id_rsa.pub hadoop@dn1
/usr/bin/ssh-copy-id: INFO: attempting to log in with the new key(s), to filter out any that are already installed
/usr/bin/ssh-copy-id: INFO: 1 key(s) remain to be installed — if you are prompted now it is to install the new keys
hadoop@dn1’s password:
Number of key(s) added: 1
Install Hadoop
Download and unzip on all nodes
hadoop@nn:~$ wget http://apache.mirror.iweb.ca/hadoop/common/hadoop-2.4.0/hadoop-2.4.0.tar.gz
hadoop@nn:~$ scp hadoop-2.4.0.tar.gz dn1:/home/hadoop/
hadoop-2.4.0.tar.gz 100% 15MB 14.6MB/s 00:00
hadoop@nn:~$ scp hadoop-2.4.0.tar.gz dn2:/home/hadoop/
hadoop-2.4.0.tar.gz 100% 15MB 14.6MB/s 00:00
change JAVA_HOME in Hadoop-env.sh
hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ vi hadoop-env.sh
#export JAVA_HOME=${JAVA_HOME}
export JAVA_HOME=/usr/local/jdk1.7.0_55
hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ scp hadoop-env.sh dn1:/home/hadoop/hadoop-2.4.0/etc/hadoop/
hadoop-env.sh 100% 3535 3.5KB/s 00:00
hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ scp hadoop-env.sh dn2:/home/hadoop/hadoop-2.4.0/etc/hadoop/
hadoop-env.sh 100% 3535 3.5KB/s 00:00
update the .bash_profile of Hadoop user
hadoop@nn:~$ more .bash_profile
export HADOOP_HOME=/home/hadoop/hadoop-2.4.0
export PATH=/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games:/usr/local/jdk1.7.0_55/bin:/home/hadoop/hadoop-2.4
.0/bin:/home/hadoop/hadoop-2.4.0/sbin
alias h=’cd /home/hadoop/hadoop-2.4.0/’
alias etc=’cd /home/hadoop/hadoop-2.4.0//etc/hadoop’
export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
export HADOOP_OPTS=”-Djava.library.path=$HADOOP_HOME/lib/native”
export CLASSPATH=$CLASSPATH:$HADOOP_HOME/share/hadoop/common:$HADOOP_HOME/share/hadoop/common/lib:\
$HADOOP_HOME/share/hadoop/yarn:$HADOOP_HOME/share/hadoop/yarn/lib:\
$HADOOP_HOME/share/hadoop/mapreduce:$HADOOP_HOME/share/hadoop/mapreduce/lib:\
$HADOOP_HOME/share/hadoop/hdfs:$HADOOP_HOME/share/hadoop/hdfs/lib
Write configurations
Core-site.xml
<configuration>
<property>
<name>fs.default.name</name>
<value>hdfs://nn:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/home/hadoop/hdfs/tmp</value>
</property>
<property>
<name>fs.checkpoint.dir</name>
<value>/home/hadoop/hdfs/checkpoint</value>
</property>
<property>
<name>hadoop.logfile.size</name>
<value>104857600</value>
</property>
<property>
<name>hadoop.logfile.count</name>
<value>20</value>
</property>
<property>
<name>io.file.buffer.size</name>
<value>131072</value>
</property>
</configuration>
Hdfs-site.xml
<configuration>
<property>
<name>dfs.name.dir</name>
<value>/home/hadoop/hdfs/name</value>
<description>Path on the local filesystem where the NameNode stores the namespace and transactions logs persistently.</description>
</property>
<property>
<name>dfs.data.dir</name>
<value>/home/hadoop/hdfs/data</value>
<description>Comma separated list of paths on the local filesystem of a DataNode where it should store its blocks.</description>
</property>
<property>
<name>dfs.replication</name>
<value>2</value>
</property>
<property>
<name>dfs.datanode.du.reserved</name>
<value>0</value>
</property>
</configuration>
Mapred-site.xml
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<property>
<name>mapreduce.map.memory.mb</name>
<value>1536</value>
</property>
<property>
<name>mapred.child.java.opts</name>
<value>-Xmx4096m</value>
</property>
<property>
<name>mapreduce.map.java.opts</name>
<value>-Xmx1228m</value>
<property>
</property>
<name>mapreduce.reduce.memory.mb</name>
<value>3072</value>
</property>
<property>
<name>mapreduce.reduce.java.opts</name>
<value>-Xmx2457m</value>
</property>
<property>
<name>mapreduce.reduce.shuffle.parallelcopies</name>
<value>50</value>
</property>
<property>
<name>mapreduce.task.io.sort.mb</name>
<value>614</value>
</property>
<property>
<name>mapreduce.task.io.sort.factor</name>
<value>100</value>
</property>
<property>
<name>mapreduce.jobhistory.address</name>
<value>nn:10020</value>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>nn:19888</value>
</property>
</configuration>
Yarn-site.xml
<configuration>
<property>
<name>yarn.resourcemanager.address</name>
<value>nn:8132</value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.address</name>
<value>nn:8130</value>
</property>
<property>
<name>yarn.resourcemanager.resource-tracker.address</name>
<value>nn:8131</value>
</property>
<property>
<name>yarn.resourcemanager.webapp.address</name>
<value>nn:8188</value>
</property>
<property>
<name>yarn.resourcemanager.amin.address</name>
<value>nn:8288</value>
</property>
<property>
<name>yarn.scheduler.minimum-allocation-mb</name>
<value>1536</value>
</property>
<property>
<name>yarn.scheduler.maximum-allocation-mb</name>
<value>43008</value>
</property>
<property>
<name>yarn.nodemanager.resource.memory-mb</name>
<value>43008</value>
</property>
<property>
<name>yarn.scheduler.minimum-allocation-vcores</name>
<value>1</value>
</property>
<property>
<name>yarn.scheduler.maximum-allocation-vcores</name>
<value>12</value>
</property>
<property>
<name>yarn.app.mapreduce.am.resource.mb</name>
<value>3072</value>
</property>
<property>
<name>yarn.app.mapreduce.am.command-opts</name>
<value>-Xmx2457m</value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.class</name>
<value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>
</property>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.nodemanager.local-dirs</name>
<value>${hadoop.tmp.dir}/nodemanager/local</value>
</property>
<property>
<name>yarn.nodemanager.address</name>
<value>0.0.0.0:8034</value>
<description>the nodemanagers bind to this port</description>
</property>
<property>
<name>yarn.nodemanager.remote-app-log-dir</name>
<value>${hadoop.tmp.dir}/nodemanager/remote</value>
<description>directory on hdfs where the application logs are moved to </description>
</property>
<property>
<name>yarn.app.mapreduce.am.command-opts</name>
<value>-Xmx2457m</value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.class</name>
<value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>
</property>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.nodemanager.local-dirs</name>
<value>${hadoop.tmp.dir}/nodemanager/local</value>
</property>
<property>
<name>yarn.nodemanager.address</name>
<value>0.0.0.0:8034</value>
<description>the nodemanagers bind to this port</description>
</property>
<property>
<name>yarn.nodemanager.remote-app-log-dir</name>
<value>${hadoop.tmp.dir}/nodemanager/remote</value>
<description>directory on hdfs where the application logs are moved to </description>
</property>
<property>
<name>yarn.nodemanager.log-dirs</name>
<value>${hadoop.tmp.dir}/nodemanager/logs</value>
<description>the directories used by Nodemanagers as log directories</description>
</property>
<property>
<name>yarn.application.classpath</name>
<value> /home/hadoop/hadoop-2.4.0/etc/hadoop,/home/hadoop/hadoop-2.4.0/share/hadoop/common/*,/home/hadoop/hadoop-2.4.0/share/hadoop/common/lib/*,/home/hadoop/hadoop-2.4.0/share/hadoop/hdfs/*,/home/hadoop/hadoop-2.4.0/share/hadoop/hdfs/lib/*,/home/hadoop/hadoop-2.4.0/share/hadoop/mapreduce/*,/home/hadoop/hadoop-2.4.0/share/hadoop/mapreduce/lib/*,/home/hadoop/hadoop-2.4.0/share/hadoop/yarn/*,/home/hadoop/hadoop-2.4.0/share/hadoop/yarn/lib/*
</value>
</property>
<property>
<name>yarn.log-aggregation-enable</name>
<value>true</value>
</property>
</configuration>
Setup slaves, master
hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ more slaves
dn1
dn2
hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ more master
nn
copy files to other nodes
copy slaves, master to dn1, dn2
hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ scp mapred-site.xml core-site.xml hdfs-site.xml yarn-site.xml dn1:/home/hadoop/hadoop-2.4.0/etc/hadoop/
mapred-site.xml 100% 2028 2.0KB/s 00:00
core-site.xml 100% 1420 1.4KB/s 00:00
hdfs-site.xml 100% 1590 1.6KB/s 00:00
yarn-site.xml 100% 5301 5.2KB/s 00:00
hadoop@nn:~/hadoop-2.4.0/etc/hadoop$ scp mapred-site.xml core-site.xml hdfs-site.xml yarn-site.xml dn2:/home/hadoop/hadoop-2.4.0/etc/hadoop/
mapred-site.xml 100% 2028 2.0KB/s 00:00
core-site.xml 100% 1420 1.4KB/s 00:00
hdfs-site.xml 100% 1590 1.6KB/s 00:00
yarn-site.xml 100% 5301 5.2KB/s 00:00
on the master, start namenode, resourcemanager and etc
136 hadoop namenode -format
137 hadoop-daemon.sh –script hdfs start namenode
138 ps -ef|grep hadoop
139 yarn-daemon.sh start resourcemanager
140 ps -ef|grep hadoop
141 yarn-daemon.sh start proxyserver
142 mr-jobhistory-daemon.sh start historyserver
On the data nodes, starts datanode, nodemanager
hadoop@dn1:~/hadoop-2.4.0/logs$ hadoop-daemon.sh –script hdfs start datanode
starting datanode, logging to /home/hadoop/hadoop-2.4.0/logs/hadoop-hadoop-datanode-dn1.out
hadoop@dn1:~/hadoop-2.4.0/logs$ yarn-daemon.sh start nodemanager
starting nodemanager, logging to /home/hadoop/hadoop-2.4.0/logs/yarn-hadoop-nodemanager-dn1.out
hadoop@dn2:~$ hadoop-daemon.sh –script hdfs start datanode
starting datanode, logging to /home/hadoop/hadoop-2.4.0/logs/hadoop-hadoop-datanode-dn2.out
hadoop@dn2:~$ yarn-daemon.sh start nodemanager
starting nodemanager, logging to /home/hadoop/hadoop-2.4.0/logs/yarn-hadoop-nodemanager-dn2.out
check status
Image may be NSFW.
Clik here to view.
hadoop@nn:~$ jps
4359 ResourceManager
6691 Jps
4275 NameNode
4652 JobHistoryServer
hadoop@dn2:~$ jps
4033 DataNode
4089 NodeManager
5392 Jps
Fix the WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform
Recompile with source codes and replace the libs
configure: error: C++ preprocessor “/lib/cpp” fails sanity check
solution: apt-get install g++
root@nn:~/Downloads/protobuf-2.5.0# protoc –version
protoc: error while loading shared libraries: libprotoc.so.8: cannot open shared object file: No such file or directory
solution: add this to /etc/profile
export LD_LIBRARY_PATH=/usr/local/lib:/lib64:/usr/lib64
root@nn:~/Downloads/protobuf-2.5.0# source /etc/profile
root@nn:~/Downloads/protobuf-2.5.0# protoc –version
apt-get install zlib1g-dev libncurses5-dev
apt-get install gcc build-essential libssl-dev zlib1g-dev libglib2.0-dev cmake maven
hadoop@nn:~/hadoop-2.4.0-src$ mvn clean package -Pdist,native -DskipTests –Dtar
…
[INFO] BUILD SUCCESS
[INFO] ————————————————————————
[INFO] Total time: 17:18.833s
[INFO] Finished at: Tue May 27 07:55:58 PDT 2014
[INFO] Final Memory: 91M/239M
[INFO] ————————————————————————
hadoop@nn:~/hadoop-2.4.0-src/hadoop-common-project/hadoop-common/target/native/target/usr/local/lib$ ls -l
total 1732
-rw-rw-r– 1 hadoop hadoop 1111258 May 27 07:39 libhadoop.a
lrwxrwxrwx 1 hadoop hadoop 18 May 27 07:39 libhadoop.so -> libhadoop.so.1.0.0
-rwxrwxr-x 1 hadoop hadoop 657597 May 27 07:39 libhadoop.so.1.0.0
hadoop@nn:~/hadoop-2.4.0-src/hadoop-common-project/hadoop-common/target/native/target/usr/local/lib$ cp * /home/hadoop/hadoop-2.4.0/lib/native/
hadoop@nn:~$ cd example
hadoop@nn:~/example$ ls
file1.txt file2.txt
hadoop@nn:~/example$ hadoop fs -mkdir /data
hadoop@nn:~/example$ hadoop fs -put -f file1.txt file2.txt /data
map reduce job failed
hadoop@nn:~/hadoop-2.4.0$ hadoop jar ./share/hadoop/mapreduce/sources/hadoop-mapreduce-examples-2.4.0-sources.jar org.apache.hadoop.examples.WordCount /data /output
14/05/27 11:30:54 INFO client.RMProxy: Connecting to ResourceManager at nn/192.168.47.139:8132
14/05/27 11:30:55 INFO input.FileInputFormat: Total input paths to process : 2
14/05/27 11:30:55 INFO mapreduce.JobSubmitter: number of splits:2
14/05/27 11:30:56 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1401214983224_0001
14/05/27 11:30:57 INFO impl.YarnClientImpl: Submitted application application_1401214983224_0001
14/05/27 11:30:57 INFO mapreduce.Job: The url to track the job: http://nn:8188/proxy/application_1401214983224_0001/
14/05/27 11:30:57 INFO mapreduce.Job: Running job: job_1401214983224_0001
14/05/27 11:31:04 INFO mapreduce.Job: Job job_1401214983224_0001 running in uber mode : false
14/05/27 11:31:04 INFO mapreduce.Job: map 0% reduce 0%
14/05/27 11:31:04 INFO mapreduce.Job: Job job_1401214983224_0001 failed with state FAILED due to: Application application_1401214983224_0001 failed 2 times due to AM Container for appattempt_1401214983224_0001_000002 exited with exitCode: 1 due to: Exception from container-launch: org.apache.hadoop.util.Shell$ExitCodeException:
org.apache.hadoop.util.Shell$ExitCodeException:
at org.apache.hadoop.util.Shell.runCommand(Shell.java:505)
at org.apache.hadoop.util.Shell.run(Shell.java:418)
at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)
at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)
at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Container exited with a non-zero exit code 1
.Failing this attempt.. Failing the application.
14/05/27 11:31:04 INFO mapreduce.Job: Counters: 0
hadoop@nn:~/hadoop-2.4.0$ hadoop jar ./share/hadoop/mapreduce/sources/hadoop-mapreduce-examples-2.4.0-sources.jar org.apache.had oop.examples.WordCount /data /output
14/05/27 11:44:19 INFO client.RMProxy: Connecting to ResourceManager at nn/192.168.47.139:8132
14/05/27 11:44:21 INFO input.FileInputFormat: Total input paths to process : 2
14/05/27 11:44:22 INFO mapreduce.JobSubmitter: number of splits:2
14/05/27 11:44:22 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1401216238271_0001
14/05/27 11:44:22 INFO impl.YarnClientImpl: Submitted application application_1401216238271_0001
14/05/27 11:44:23 INFO mapreduce.Job: The url to track the job: http://nn:8188/proxy/application_1401216238271_0001/
14/05/27 11:44:23 INFO mapreduce.Job: Running job: job_1401216238271_0001
14/05/27 11:44:49 INFO mapreduce.Job: Job job_1401216238271_0001 running in uber mode : false
14/05/27 11:44:49 INFO mapreduce.Job: map 0% reduce 0%
14/05/27 11:44:58 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000000_0, Status : FAILED
Container [pid=7385,containerID=container_1401216238271_0001_01_000002] is running beyond virtual memory limits. Current usage: 67.2 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.
Dump of the process-tree for container_1401216238271_0001_01_000002 :
|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE
|- 7396 7385 7385 7385 (java) 179 18 4726177792 16837 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -D hadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/appli cation_1401216238271_0001/container_1401216238271_0001_01_000002/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app .container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000002 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attemp t_1401216238271_0001_m_000000_0 2
|- 7385 7224 7385 7385 (bash) 0 0 17039360 354 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000002/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000002 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000000_0 2 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000002/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000002/stderr
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143
14/05/27 11:45:00 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000001_0, Status : FAILED
Container [pid=7384,containerID=container_1401216238271_0001_01_000003] is running beyond virtual memory limits. Current usage: 68.6 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.
Dump of the process-tree for container_1401216238271_0001_01_000003 :
|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE
|- 7394 7384 7384 7384 (java) 178 18 4726177792 17213 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -D hadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/appli cation_1401216238271_0001/container_1401216238271_0001_01_000003/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app .container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000003 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attemp t_1401216238271_0001_m_000001_0 3
|- 7384 7224 7384 7384 (bash) 0 0 17039360 354 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000003/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000003 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000001_0 3 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000003/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000003/stderr
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143
14/05/27 11:45:08 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000000_1, Status : FAILED
Container [pid=7441,containerID=container_1401216238271_0001_01_000004] is running beyond virtual memory limits. Current usage: 59.4 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.
Dump of the process-tree for container_1401216238271_0001_01_000004 :
|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE
|- 7441 7224 7441 7441 (bash) 0 0 17039360 354 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000004/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000004 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000000_1 4 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000004/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000004/stderr
|- 7446 7441 7441 7441 (java) 178 8 4726177792 14843 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -Dh adoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/applic ation_1401216238271_0001/container_1401216238271_0001_01_000004/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app. container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000004 – Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt _1401216238271_0001_m_000000_1 4
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143
14/05/27 11:45:08 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000001_1, Status : FAILED
Container [pid=7456,containerID=container_1401216238271_0001_01_000005] is running beyond virtual memory limits. Current usage: 58.4 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.
Dump of the process-tree for container_1401216238271_0001_01_000005 :
|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE
|- 7461 7456 7456 7456 (java) 177 9 4726177792 14603 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -Dh adoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/applic ation_1401216238271_0001/container_1401216238271_0001_01_000005/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app. container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000005 – Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt _1401216238271_0001_m_000001_1 5
|- 7456 7224 7456 7456 (bash) 0 0 17039360 353 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000005/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000005 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000001_1 5 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000005/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000005/stderr
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143
14/05/27 11:45:17 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000000_2, Status : FAILED
Container [pid=7089,containerID=container_1401216238271_0001_01_000008] is running beyond virtual memory limits. Current usage: 63.2 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.
Dump of the process-tree for container_1401216238271_0001_01_000008 :
|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE
|- 7089 6803 7089 7089 (bash) 0 1 17039360 353 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000008/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000008 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000000_2 8 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000008/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000008/stderr
|- 7095 7089 7089 7089 (java) 182 25 4726177792 15832 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -D hadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/appli cation_1401216238271_0001/container_1401216238271_0001_01_000008/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app .container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000008 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attemp t_1401216238271_0001_m_000000_2 8
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143
14/05/27 11:45:17 INFO mapreduce.Job: Task Id : attempt_1401216238271_0001_m_000001_2, Status : FAILED
Container [pid=7107,containerID=container_1401216238271_0001_01_000009] is running beyond virtual memory limits. Current usage: 54.7 MB of 1.5 GB physical memory used; 4.4 GB of 3.1 GB virtual memory used. Killing container.
Dump of the process-tree for container_1401216238271_0001_01_000009 :
|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL _CMD_LINE
|- 7113 7107 7107 7107 (java) 180 19 4726177792 13644 /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=true -D hadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcache/appli cation_1401216238271_0001/container_1401216238271_0001_01_000009/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app .container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_000009 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attemp t_1401216238271_0001_m_000001_2 9
|- 7107 6803 7107 7107 (bash) 0 0 17039360 353 /bin/bash -c /usr/local/jdk1.7.0_55/bin/java -Djava.net.preferIPv4Stack=t rue -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/hdfs/tmp/nodemanager/local/usercache/hadoop/appcach e/application_1401216238271_0001/container_1401216238271_0001_01_000009/tmp -Dlog4j.configuration=container-log4j.properties -Dy arn.app.container.log.dir=/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0001_01_ 000009 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.47.141 37535 attempt_1401216238271_0001_m_000001_2 9 1>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_14012 16238271_0001_01_000009/stdout 2>/home/hadoop/hdfs/tmp/nodemanager/logs/application_1401216238271_0001/container_1401216238271_0 001_01_000009/stderr
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143
14/05/27 11:45:24 INFO mapreduce.Job: map 100% reduce 100%
14/05/27 11:45:24 INFO mapreduce.Job: Job job_1401216238271_0001 failed with state FAILED due to: Task failed task_1401216238271 _0001_m_000001
Job failed as tasks failed. failedMaps:1 failedReduces:0
14/05/27 11:45:24 INFO mapreduce.Job: Counters: 13
Job Counters
Failed map tasks=7
Killed map tasks=1
Launched map tasks=8
Other local map tasks=6
Data-local map tasks=2
Total time spent by all maps in occupied slots (ms)=50333
Total time spent by all reduces in occupied slots (ms)=0
Total time spent by all map tasks (ms)=50333
Total vcore-seconds taken by all map tasks=50333
Total megabyte-seconds taken by all map tasks=77311488
Map-Reduce Framework
CPU time spent (ms)=0
Physical memory (bytes) snapshot=0
Virtual memory (bytes) snapshot=0
hadoop@nn:~/hadoop-2.4.0$ hadoop jar ./share/hadoop/mapreduce/sources/hadoop-mapreduce-examples-2.4.0-sources.jar org.apache.hadoop.examples.WordCount /data /output4
14/05/27 13:16:57 INFO client.RMProxy: Connecting to ResourceManager at nn/192.168.47.139:8132
14/05/27 13:16:58 INFO input.FileInputFormat: Total input paths to process : 2
14/05/27 13:16:58 INFO mapreduce.JobSubmitter: number of splits:2
14/05/27 13:16:59 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1401221062327_0001
14/05/27 13:16:59 INFO impl.YarnClientImpl: Submitted application application_1401221062327_0001
14/05/27 13:16:59 INFO mapreduce.Job: The url to track the job: http://nn:8188/proxy/application_1401221062327_0001/
14/05/27 13:16:59 INFO mapreduce.Job: Running job: job_1401221062327_0001
14/05/27 13:17:26 INFO mapreduce.Job: Job job_1401221062327_0001 running in uber mode : false
14/05/27 13:17:26 INFO mapreduce.Job: map 0% reduce 0%
14/05/27 13:18:01 INFO mapreduce.Job: map 100% reduce 0%
14/05/27 13:18:30 INFO mapreduce.Job: map 100% reduce 100%
14/05/27 13:18:30 INFO mapreduce.Job: Job job_1401221062327_0001 completed successfully
14/05/27 13:18:30 INFO mapreduce.Job: Counters: 49
File System Counters
FILE: Number of bytes read=103
FILE: Number of bytes written=282013
FILE: Number of read operations=0
FILE: Number of large read operations=0
FILE: Number of write operations=0
HDFS: Number of bytes read=272
HDFS: Number of bytes written=56
HDFS: Number of read operations=9
HDFS: Number of large read operations=0
HDFS: Number of write operations=2
Job Counters
Launched map tasks=2
Launched reduce tasks=1
Data-local map tasks=2
Total time spent by all maps in occupied slots (ms)=65907
Total time spent by all reduces in occupied slots (ms)=25452
Total time spent by all map tasks (ms)=65907
Total time spent by all reduce tasks (ms)=25452
Total vcore-seconds taken by all map tasks=65907
Total vcore-seconds taken by all reduce tasks=25452
Total megabyte-seconds taken by all map tasks=67488768
Total megabyte-seconds taken by all reduce tasks=26062848
Map-Reduce Framework
Map input records=6
Map output records=12
Map output bytes=123
Map output materialized bytes=109
Input split bytes=188
Combine input records=12
Combine output records=8
Reduce input groups=7
Reduce shuffle bytes=109
Reduce input records=8
Reduce output records=7
Spilled Records=16
Shuffled Maps =2
Failed Shuffles=0
Merged Map outputs=2
GC time elapsed (ms)=424
CPU time spent (ms)=3410
Physical memory (bytes) snapshot=486014976
Virtual memory (bytes) snapshot=6566645760
Total committed heap usage (bytes)=256843776
Shuffle Errors
BAD_ID=0
CONNECTION=0
IO_ERROR=0
WRONG_LENGTH=0
WRONG_MAP=0
WRONG_REDUCE=0
File Input Format Counters
Bytes Read=84
File Output Format Counters
Bytes Written=56
hadoop@nn:~/hadoop-2.4.0$ hadoop fs -rm -r /output
14/05/27 16:38:41 INFO fs.TrashPolicyDefault: Namenode trash configuration: Deletion interval = 0 minutes, Emptier interval = 0 minutes.
Deleted /output
The post Multinodes Hadoop 2.4 installation and configuration appeared first on Robert Chen.