mapreduce job failed with alluxio configuration

102 views
Skip to first unread message

Kaiming Wan

unread,
Sep 29, 2016, 9:27:22 AM9/29/16
to Alluxio Users
I ran a MR job on my cluster on 3 nodes which have 128GB and 40 cores each.

Cluster Info:
10.8.12.16: namenode active,datanode,alluxio master,alluxio worker
10.8.12.17: namenode standby,datanode,alluxio master standby,alluxio worker
10.8.12.18: datanode,alluxio worker


version info:
os: red hat el6
alluxio: 1.2.0
hadoop:2.7.2
jdk:1.8


I have compiled alluxio with hadoop version 2.7.2 



I have configured hdfs and alluxio with HA. And both alluxio and hdfs can start successfully without any error in logs.

When just ran my mapreduce job without using alluxio, it failed with error:

16/09/29 21:10:30 INFO mapreduce.Job: Job job_1475154019111_0005 failed with state FAILED due to: Application application_1475154019111_0005 failed 2 times due to AM Container for appattempt_1475154019111_0005_000002 exited with  exitCode: -1
For more detailed output, check application tracking page:http://sq-hbase1.800best.com:8088/cluster/app/application_1475154019111_0005Then, click on links to logs of each attempt.
Diagnostics: org.apache.commons.codec.binary.Base64.encodeBase64String([B)Ljava/lang/String;
Failing this attempt. Failing the application.
16/09/29 21:10:30 INFO mapreduce.Job: Counters: 0


My MR line count job source code:
/**
 * MR统计一个文件的行数
 * @author Wan Kaiming on 2016/9/9
 * @version 1.0
 */

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;


import java.io.IOException;
public class MRLineCount {




   
//Mapper主要就是每行形成一个分片,key是一行的内容,value是值1
   
public static class LineMapper
           
extends Mapper<Object, Text, Text, IntWritable>{


       
//统计行数
       
private final static IntWritable one = new IntWritable(1);
       
private Text word = new Text("总行数为:");


       
public void map(Object key, Text value, Context context
       
) throws IOException, InterruptedException {


           
//由于我们使用TextInputFormar,其中Key是每个数据记录在数据分片中字节偏移量 Text类型value存储的是一行的内容


           
//传递给reducer的key全部固定为一个值,value就是值1,代表一行。这样reducer可以全部在一个key里面求和
            context
.write(word, one);


       
}
   
}




   
//reducer对行数进行统计求和,注意输出的key为null,我们只要计算总数即可
   
public static class IntSumReducer
           
extends Reducer<Text,IntWritable,Text,IntWritable> {
       
private IntWritable result = new IntWritable();


       
public void reduce(Text key, Iterable<IntWritable> values,
                           
Context context
       
) throws IOException, InterruptedException {
           
int sum = 0;
           
for (IntWritable val : values) {
                sum
+= val.get();
           
}
            result
.set(sum);
            context
.write(key, result);
       
}
   
}


   
public static void main(String[] args) throws Exception {
       
System.setProperty("HADOOP_USER_NAME", "appadmin");


       
Configuration conf = new Configuration();


       
//设置hdfs和yarn地址
       
//使用alluxio
       
//conf.set("fs.defaultFS", "alluxio://10.8.12.16:19998");
       
//使用hdfs
        conf
.set("fs.defaultFS", "hdfs://10.8.12.17:9000");
        conf
.set("yarn.resourcemanager.hostname","10.8.12.16");


       
System.out.println(" codec location "+org.apache.commons.codec.binary.Base64.class.getProtectionDomain().getCodeSource());


       
//使用默认配置创建一个Job实例
       
Job job = Job.getInstance(conf, "linecount");


        job
.setJar("E:\\JavaProjects\\Learning\\out\\artifacts\\hadoop_test_jar\\hadoop-test.jar");
        job
.setJarByClass(MRLineCount.class);


       
//设置mapper,combiner和reducer
        job
.setMapperClass(LineMapper.class);
        job
.setCombinerClass(IntSumReducer.class);
        job
.setReducerClass(IntSumReducer.class);
       
//设置输出的key,value类,由于我们不需要输出key的内容,就使用NullWritable类型
        job
.setOutputKeyClass(Text.class);
        job
.setOutputValueClass(IntWritable.class);
       
//设置参数
       
FileInputFormat.addInputPath(job, new Path(args[0]));
       
FileOutputFormat.setOutputPath(job, new Path(args[1]));


       
System.exit(job.waitForCompletion(true) ? 0 : 1);
   
}
}



hdfs-site.xml
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
  Licensed under the Apache License, Version 2.0 (the "License");
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at


    http://www.apache.org/licenses/LICENSE-2.0


  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an "AS IS" BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
-->



<!-- Put site-specific property overrides in this file. -->






<configuration>
 
<!--复制的份数,默认为3份,仅仅在datanode上配置-->
       
<property>
       
<name>dfs.replication</name>
               
<value>2</value>
       
</property>


 
<!--存储命名空间和操作日志相关的元数据信息的本地文件系统目录-->
       
<property>
       
<name>dfs.namenode.name.dir</name>
               
<value>file:/home/appadmin/hadoop-2.7.2/hdfs/namenode</value>
       
</property>


<!--存储命名空间和操作日志相关的元数据信息的本地文件系统目录-->
       
<property>
               
<name>dfs.namenode.data.dir</name>
               
<value>file:/home/appadmin/hadoop-2.7.2/hdfs/datanode</value>
       
</property>


 
<!--指定hdfs的nameservice为ns,需要和core-site.xml中的保持一致 -->    
   
<property>    
           
<name>dfs.nameservices</name>    
           
<value>ns</value>    
   
</property>  
   
<!-- ns下面有两个NameNode,分别是nn1,nn2,名字可以自取,最多配置2个 -->
   
<property>
           
<name>dfs.ha.namenodes.ns</name>
           
<value>nn1,nn2</value>
   
</property>
                 


<!-- nn1的RPC通信地址 -->
   
<property>
           
<name>dfs.namenode.rpc-address.ns.nn1</name>
           
<value>10.8.12.16:9000</value>
   
</property>
     
<!-- nn1的http通信地址 -->
   
<property>
           
<name>dfs.namenode.http-address.ns.nn1</name>
           
<value>10.8.12.16:50070</value>
   
</property>
                           
<!-- nn2的RPC通信地址 -->
   
<property>
           
<name>dfs.namenode.rpc-address.ns.nn2</name>
           
<value>10.8.12.17:9000</value>
   
</property>
                           
<!-- nn2的http通信地址 -->
   
<property>
           
<name>dfs.namenode.http-address.ns.nn2</name>
           
<value>10.8.12.17:50070</value>
   
</property>
                           
<!-- 指定NameNode的元数据(edit log)存放在哪些JournalNode上,以及存放位置,最后的是journalid,一般用集群的nameservice名字来代替 -->
   
<property>
             
<name>dfs.namenode.shared.edits.dir</name>
             
<value>qjournal://10.8.12.16:8485;10.8.12.17:8485;10.8.12.18:8485/ns</value>
   
</property>
     
                 
<!-- 指定JournalNode在本地磁盘存放数据的位置 -->
   
<property>
             
<name>dfs.journalnode.edits.dir</name>
             
<value>/home/appadmin/hadoop-2.7.2/journal</value>
   
</property>
                               
<!-- 开启NameNode故障时自动切换 -->
   
<property>
             
<name>dfs.ha.automatic-failover.enabled</name>
             
<value>true</value>
   
</property>
                               
<!-- 配置失败自动切换实现方式 -->
   
<property>
               
<name>dfs.client.failover.proxy.provider.ns</name>
               
<value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
   
</property>
                                   
<!-- 配置隔离机制 -->
<!--
这个参数比较重要,主要用于在主备节点切换时实现隔离机制的,在官方网站中做了相当详细的配置说明,其大概意思为:主备架构解决单点故障问题时,必须要认真解决的是脑裂问题,即出现两个 master 同时对外提供服务,导致系统处于不一致状态,可能导致数据丢失等潜在问题。在 HDFS HA 中, JournalNode 只允许一个 NameNode 写数据,不会出现两个 Active NameNode 的问题,但是,当主备切换时,之前的 Active NameNode 可能仍在处理客户端的 RPC 请求,为此,需要增加隔离机制( fencing )将之前的 Active NameNode 杀死。 HDFS 允许用户配置多个隔离机制,当发生主备切换时,将顺次执行这些隔离机制,直到一个返回成功。 Hadoop 2.0 内部打包了两种类型的隔离机制,分别是 shell  和 sshfence 。


-->

   
<property>
                 
<name>dfs.ha.fencing.methods</name>
                 
<value>sshfence</value>
   
</property>
                                     


<!-- 使用隔离机制时需要ssh免登陆 -->
   
<property>
               
<name>dfs.ha.fencing.ssh.private-key-files</name>
               
<value>/home/appadmin/.ssh/id_rsa</value>
   
</property>
                                                             
                           
                         
<!-- 在NN和DN上开启WebHDFS (REST API)功能,不是必须 -->                                                                    
   
<property>    
           
<name>dfs.webhdfs.enabled</name>    
           
<value>true</value>    
   
</property>    




 
<property>
 
<name>fs.alluxio.impl</name>
   
<value>alluxio.hadoop.FileSystem</value>
     
<description>The Alluxio FileSystem (Hadoop 1.x and 2.x)</description>
     
</property>
     
<property>
       
<name>fs.alluxio-ft.impl</name>
         
<value>alluxio.hadoop.FaultTolerantFileSystem</value>
         
<description>The Alluxio FileSystem (Hadoop 1.x and 2.x) with fault tolerant support</description>
       
</property>
       
<property>
             
<name>fs.AbstractFileSystem.alluxio.impl</name>
               
<value>alluxio.hadoop.AlluxioFileSystem</value>
               
<description>The Alluxio AbstractFileSystem (Hadoop 2.x)</description>
       
</property>






</configuration>


core-site.xml
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
  Licensed under the Apache License, Version 2.0 (the "License");
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at


    http://www.apache.org/licenses/LICENSE-2.0


  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an "AS IS" BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
-->



<!-- Put site-specific property overrides in this file. -->


<configuration>
 
<property>
 
<name>hadoop.tmp.dir</name>
 
<value>/home/appadmin/hadoop-2.7.2/tmp</value>
 
</property>
 
<!--高可用配置,这里使用hsfs-site.xml中配置的nameservice-->
 
<property>
 
<name>fs.defaultFS</name>
 
<value>hdfs://ns</value>
 
</property>
 
<!--使用zk用于主备namenode切换-->
 
<property>
       
<name>ha.zookeeper.quorum</name>
       
<value>10.8.12.16:2181,10.8.12.17:2181,10.8.12.18:2181</value>
 
</property>




 
<property>
 
<name>fs.alluxio.impl</name>
   
<value>alluxio.hadoop.FileSystem</value>
     
<description>The Alluxio FileSystem (Hadoop 1.x and 2.x)</description>
     
</property>
     
<property>
       
<name>fs.alluxio-ft.impl</name>
         
<value>alluxio.hadoop.FaultTolerantFileSystem</value>
       
<description>The Alluxio FileSystem (Hadoop 1.x and 2.x) with fault tolerant support</description>
       
</property>
       
<property>
             
<name>fs.AbstractFileSystem.alluxio.impl</name>
               
<value>alluxio.hadoop.AlluxioFileSystem</value>
               
<description>The Alluxio AbstractFileSystem (Hadoop 2.x)</description>
       
</property>


</configuration>



yarn-site.xml
<configuration>


 
<property>
 
<name>yarn.resourcemanager.address</name>
 
<value>10.8.12.16:8032</value>
 
</property>
 
<property>
 
<name>yarn.resourcemanager.scheduler.address</name>
 
<value>10.8.12.16:8030</value>
 
</property>
 
<property>
 
<name>yarn.resourcemanager.resource-tracker.address</name>
 
<value>10.8.12.16:8031</value>
 
</property>
 
<property>
 
<name>yarn.resourcemanager.admin.address</name>
 
<value>10.8.12.16:8033</value>
 
</property>
 
<property>
 
<name>yarn.resourcemanager.webapp.address</name>
 
<value>10.8.12.16:8088</value>
 
</property>
 
<property>
 
<name>mapreduce.framework.name</name>
 
<value>yarn</value>
 
</property>
 
<!--支持mr任务提交到yarn-->
 
<property>
 
<name>yarn.nodemanager.aux-services</name>
 
<value>mapreduce_shuffle</value>
 
</property>


 
 
<property>
 
<name>yarn.nodemanager.resource.memory-mb</name>
 
<value>81408</value>
 
</property>


 
<property>
 
<name>yarn.scheduler.minimum-allocation-mb</name>
 
<value>27136</value>
 
</property>


 
<property>
 
<name>yarn.scheduler.maximum-allocation-mb</name>
 
<value>81408</value>
 
</property>


 
<property>
 
<name>yarn.app.mapreduce.am.resource.mb</name>
 
<value>27136</value>
 
</property>


 
<property>
 
<name>yarn.app.mapreduce.am.command-opts</name>
 
<value>-Xmx21708m</value>
 
</property>




 
<property>
 
<name>yarn.nodemanager.pmem-check-enabled</name>
 
<value>false</value>
 
</property>


 
<property>
 
<name>yarn.nodemanager.vmem-check-enabled</name>
 
<value>false</value>
 
</property>


</configuration>

mapred-site.xml
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
  Licensed under the Apache License, Version 2.0 (the "License");
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at


    http://www.apache.org/licenses/LICENSE-2.0


  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an "AS IS" BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
-->



<!-- Put site-specific property overrides in this file. -->


<configuration>
 
<property>
 
<name>mapreduce.framework.name</name>
 
<value>yarn</value>
 
</property>




 
<property>
 
<name>mapreduce.map.memory.mb</name>
 
<value>27136</value>
 
</property>


 
<property>
 
<name>mapreduce.map.java.opts</name>
 
<value>-Xmx21708m</value>
 
</property>


 
 
<property>
 
<name>mapreduce.reduce.memory.mb</name>
 
<value>27136</value>
 
</property>


 
<property>
 
<name>mapreduce.reduce.java.opts</name>
 
<value>-Xmx21708m</value>
 
</property>
</configuration>



hadoop-env.sh
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements.  See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership.  The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License.  You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.


# Set Hadoop-specific environment variables here.


# The only required environment variable is JAVA_HOME.  All others are
# optional.  When running a distributed configuration it is best to
# set JAVA_HOME in this file, so that it is correctly defined on
# remote nodes.


# The java implementation to use.
export JAVA_HOME=${JAVA_HOME}


# The jsvc implementation to use. Jsvc is required to run secure datanodes
# that bind to privileged ports to provide authentication of data transfer
# protocol.  Jsvc is not required if SASL is configured for authentication of
# data transfer protocol using non-privileged ports.
#export JSVC_HOME=${JSVC_HOME}


export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/etc/hadoop"}


# Extra Java CLASSPATH elements.  Automatically insert capacity-scheduler.
for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do
  if [ "$HADOOP_CLASSPATH" ]; then
    export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f
  else
    export HADOOP_CLASSPATH=$f
  fi
done


# The maximum amount of heap to use, in MB. Default is 1000.
#export HADOOP_HEAPSIZE=
#export HADOOP_NAMENODE_INIT_HEAPSIZE=""


# Extra Java runtime options.  Empty by default.
export HADOOP_OPTS="$HADOOP_OPTS -Djava.net.preferIPv4Stack=true"


# Command specific options appended to HADOOP_OPTS when specified
export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_NAMENODE_OPTS"
export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS $HADOOP_DATANODE_OPTS"


export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_SECONDARYNAMENODE_OPTS"


export HADOOP_NFS3_OPTS="$HADOOP_NFS3_OPTS"
export HADOOP_PORTMAP_OPTS="-Xmx512m $HADOOP_PORTMAP_OPTS"


# The following applies to multiple commands (fs, dfs, fsck, distcp etc)
export HADOOP_CLIENT_OPTS="-Xmx512m $HADOOP_CLIENT_OPTS"
#HADOOP_JAVA_PLATFORM_OPTS="-XX:-UsePerfData $HADOOP_JAVA_PLATFORM_OPTS"


# On secure datanodes, user to run the datanode as after dropping privileges.
# This **MUST** be uncommented to enable secure HDFS if using privileged ports
# to provide authentication of data transfer protocol.  This **MUST NOT** be
# defined if SASL is configured for authentication of data transfer protocol
# using non-privileged ports.
export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER}


# Where log files are stored.  $HADOOP_HOME/logs by default.
#export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER


# Where log files are stored in the secure data environment.
export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}


###
# HDFS Mover specific parameters
###
# Specify the JVM options to be used when starting the HDFS Mover.
# These options will be appended to the options specified as HADOOP_OPTS
# and therefore may override any similar flags set in HADOOP_OPTS
#
# export HADOOP_MOVER_OPTS=""


###
# Advanced Users Only!
###


# The directory where pid files are stored. /tmp by default.
# NOTE: this should be set to a directory that can only be written to by
#       the user that will run the hadoop daemons.  Otherwise there is the
#       potential for a symlink attack.
export HADOOP_PID_DIR=${HADOOP_HOME}/pids
export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR}


# A string representing this instance of hadoop. $USER by default.
export HADOOP_IDENT_STRING=$USER


export HADOOP_CLASSPATH=/home/appadmin/alluxio-1.2.0/core/client/target/alluxio-core-client-1.2.0-jar-with-dependencies.jar:${HADOOP_CLASSPATH}


PS: I have copied alluxio-core-client-1.2.0-jar-with-dependencies to $HADOOP_HOME/share/hadoop/common/lib on every node.



The yarn app job log:
2016-09-29 21:24:56,125 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Created MRAppMaster for application appattempt_1475155149952_0002_000001 2016-09-29 21:24:56,227 WARN [main] org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2016-09-29 21:24:56,252 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Executing with tokens: 2016-09-29 21:24:56,323 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Kind: YARN_AM_RM_TOKEN, Service: , Ident: (appAttemptId { application_id { id: 2 cluster_timestamp: 1475155149952 } attemptId: 1 } keyId: -1453022468) 2016-09-29 21:24:56,333 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Using mapred newApiCommitter. 2016-09-29 21:24:56,738 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter set in config null 2016-09-29 21:24:56,771 INFO [main] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: File Output Committer Algorithm version is 1 2016-09-29 21:24:56,774 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter is org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter 2016-09-29 21:24:56,787 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.jobhistory.EventType for class org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler 2016-09-29 21:24:56,788 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.JobEventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobEventDispatcher 2016-09-29 21:24:56,788 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.TaskEventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskEventDispatcher 2016-09-29 21:24:56,788 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskAttemptEventDispatcher 2016-09-29 21:24:56,789 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventType for class org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler 2016-09-29 21:24:56,792 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.speculate.Speculator$EventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$SpeculatorEventDispatcher 2016-09-29 21:24:56,792 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.rm.ContainerAllocator$EventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter 2016-09-29 21:24:56,793 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncher$EventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerLauncherRouter 2016-09-29 21:24:56,818 INFO [main] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://ns:8020] 2016-09-29 21:24:56,833 INFO [main] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://ns:8020] 2016-09-29 21:24:56,846 INFO [main] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://ns:8020] 2016-09-29 21:24:56,859 INFO [main] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Creating intermediate history logDir: [hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate] + based on conf. Should ideally be created by the JobHistoryServer: yarn.app.mapreduce.am.create-intermediate-jh-base-dir 2016-09-29 21:24:56,866 INFO [main] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Perms after creating 493, Expected: 1023 2016-09-29 21:24:56,866 INFO [main] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Explicitly setting permissions to : 1023, rwxrwxrwt 2016-09-29 21:24:56,874 INFO [main] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Perms after creating 488, Expected: 504 2016-09-29 21:24:56,874 INFO [main] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Explicitly setting permissions to : 504, rwxrwx--- 2016-09-29 21:24:56,876 INFO [main] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Emitting job history data to the timeline server is not enabled 2016-09-29 21:24:56,899 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.JobFinishEvent$Type for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobFinishEventHandler 2016-09-29 21:24:57,036 INFO [main] org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties 2016-09-29 21:24:57,078 INFO [main] org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2016-09-29 21:24:57,078 INFO [main] org.apache.hadoop.metrics2.impl.MetricsSystemImpl: MRAppMaster metrics system started 2016-09-29 21:24:57,083 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Adding job token for job_1475155149952_0002 to jobTokenSecretManager 2016-09-29 21:24:57,153 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Not uberizing job_1475155149952_0002 because: not enabled; 2016-09-29 21:24:57,163 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Input size for job job_1475155149952_0002 = 12. Number of splits = 1 2016-09-29 21:24:57,163 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Number of reduces for job job_1475155149952_0002 = 1 2016-09-29 21:24:57,164 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1475155149952_0002Job Transitioned from NEW to INITED 2016-09-29 21:24:57,164 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: MRAppMaster launching normal, non-uberized, multi-container job job_1475155149952_0002. 2016-09-29 21:24:57,179 INFO [main] org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue 2016-09-29 21:24:57,185 INFO [Socket Reader #1 for port 8540] org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 8540 2016-09-29 21:24:57,196 INFO [main] org.apache.hadoop.yarn.factories.impl.pb.RpcServerFactoryPBImpl: Adding protocol org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB to the server 2016-09-29 21:24:57,197 INFO [IPC Server Responder] org.apache.hadoop.ipc.Server: IPC Server Responder: starting 2016-09-29 21:24:57,197 INFO [IPC Server listener on 8540] org.apache.hadoop.ipc.Server: IPC Server listener on 8540: starting 2016-09-29 21:24:57,198 INFO [main] org.apache.hadoop.mapreduce.v2.app.client.MRClientService: Instantiated MRClientService at sq-hbase2.800best.com/10.8.12.17:8540 2016-09-29 21:24:57,246 INFO [main] org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 2016-09-29 21:24:57,251 INFO [main] org.apache.hadoop.security.authentication.server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2016-09-29 21:24:57,254 WARN [main] org.apache.hadoop.http.HttpRequestLog: Jetty request log can only be enabled using Log4j 2016-09-29 21:24:57,258 INFO [main] org.apache.hadoop.http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2016-09-29 21:24:57,261 INFO [main] org.apache.hadoop.http.HttpServer2: Added filter AM_PROXY_FILTER (class=org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter) to context mapreduce 2016-09-29 21:24:57,261 INFO [main] org.apache.hadoop.http.HttpServer2: Added filter AM_PROXY_FILTER (class=org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter) to context static 2016-09-29 21:24:57,263 INFO [main] org.apache.hadoop.http.HttpServer2: adding path spec: /mapreduce/* 2016-09-29 21:24:57,263 INFO [main] org.apache.hadoop.http.HttpServer2: adding path spec: /ws/* 2016-09-29 21:24:57,459 INFO [main] org.apache.hadoop.yarn.webapp.WebApps: Registered webapp guice modules 2016-09-29 21:24:57,460 INFO [main] org.apache.hadoop.http.HttpServer2: Jetty bound to port 2252 2016-09-29 21:24:57,461 INFO [main] org.mortbay.log: jetty-6.1.26 2016-09-29 21:24:57,499 INFO [main] org.mortbay.log: Extract jar:file:/home/appadmin/hadoop-2.7.2/share/hadoop/common/lib/alluxio-core-client-1.2.0-jar-with-dependencies.jar!/webapps/mapreduce to /home/appadmin/hadoop-2.7.2/tmp/nm-local-dir/usercache/appadmin/appcache/application_1475155149952_0002/container_1475155149952_0002_01_000001/tmp/Jetty_0_0_0_0_2252_mapreduce____.219gm6/webapp 2016-09-29 21:24:59,124 INFO [main] org.mortbay.log: Started HttpServer2$SelectChannelConne...@0.0.0.0:2252 2016-09-29 21:24:59,124 INFO [main] org.apache.hadoop.yarn.webapp.WebApps: Web app mapreduce started at 2252 2016-09-29 21:24:59,125 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator: JOB_CREATE job_1475155149952_0002 2016-09-29 21:24:59,126 INFO [main] org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue 2016-09-29 21:24:59,126 INFO [Socket Reader #1 for port 26213] org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 26213 2016-09-29 21:24:59,128 INFO [IPC Server Responder] org.apache.hadoop.ipc.Server: IPC Server Responder: starting 2016-09-29 21:24:59,128 INFO [IPC Server listener on 26213] org.apache.hadoop.ipc.Server: IPC Server listener on 26213: starting 2016-09-29 21:24:59,142 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: nodeBlacklistingEnabled:true 2016-09-29 21:24:59,142 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: maxTaskFailuresPerNode is 3 2016-09-29 21:24:59,142 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: blacklistDisablePercent is 33 2016-09-29 21:24:59,167 INFO [main] org.apache.hadoop.yarn.client.RMProxy: Connecting to ResourceManager at /10.8.12.16:8030 2016-09-29 21:24:59,217 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator: maxContainerCapability: <memory:81408, vCores:32> 2016-09-29 21:24:59,217 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator: queue: default 2016-09-29 21:24:59,219 INFO [main] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Upper limit on the thread pool size is 500 2016-09-29 21:24:59,219 INFO [main] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: The thread pool initial size is 10 2016-09-29 21:24:59,221 INFO [main] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: yarn.client.max-cached-nodemanagers-proxies : 0 2016-09-29 21:24:59,224 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1475155149952_0002Job Transitioned from INITED to SETUP 2016-09-29 21:24:59,225 INFO [CommitterEvent Processor #0] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: JOB_SETUP 2016-09-29 21:24:59,230 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1475155149952_0002Job Transitioned from SETUP to RUNNING 2016-09-29 21:24:59,239 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved sq-hbase1.800best.com to /default-rack 2016-09-29 21:24:59,239 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved sq-hbase2.800best.com to /default-rack 2016-09-29 21:24:59,240 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1475155149952_0002_m_000000 Task Transitioned from NEW to SCHEDULED 2016-09-29 21:24:59,241 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1475155149952_0002_r_000000 Task Transitioned from NEW to SCHEDULED 2016-09-29 21:24:59,241 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_m_000000_0 TaskAttempt Transitioned from NEW to UNASSIGNED 2016-09-29 21:24:59,242 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_0 TaskAttempt Transitioned from NEW to UNASSIGNED 2016-09-29 21:24:59,242 INFO [Thread-51] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: mapResourceRequest:<memory:1024, vCores:1> 2016-09-29 21:24:59,246 INFO [Thread-51] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: reduceResourceRequest:<memory:1024, vCores:1> 2016-09-29 21:24:59,274 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Event Writer setup for JobId: job_1475155149952_0002, File: hdfs://ns:8020/tmp/hadoop-yarn/staging/appadmin/.staging/job_1475155149952_0002/job_1475155149952_0002_1.jhist 2016-09-29 21:25:00,219 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:0 ContRel:0 HostLocal:0 RackLocal:0 2016-09-29 21:25:00,244 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1475155149952_0002: ask=4 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:217088, vCores:1> knownNMs=3 2016-09-29 21:25:00,244 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:217088, vCores:1> 2016-09-29 21:25:00,245 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 1 2016-09-29 21:25:01,254 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 1 2016-09-29 21:25:01,255 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1475155149952_0002_01_000002 to attempt_1475155149952_0002_m_000000_0 2016-09-29 21:25:01,256 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:189952, vCores:1> 2016-09-29 21:25:01,256 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 1 2016-09-29 21:25:01,256 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:01,280 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved sq-hbase2.800best.com to /default-rack 2016-09-29 21:25:01,290 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: The job-jar file on the remote FS is hdfs://10.8.12.17:9000/tmp/hadoop-yarn/staging/appadmin/.staging/job_1475155149952_0002/job.jar 2016-09-29 21:25:01,292 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: The job-conf file on the remote FS is /tmp/hadoop-yarn/staging/appadmin/.staging/job_1475155149952_0002/job.xml 2016-09-29 21:25:01,293 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Adding #0 tokens and #1 secret keys for NM use for launching container 2016-09-29 21:25:01,293 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Size of containertokens_dob is 1 2016-09-29 21:25:01,293 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Putting shuffle token in serviceData 2016-09-29 21:25:01,307 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_m_000000_0 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED 2016-09-29 21:25:01,310 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1475155149952_0002_01_000002 taskAttempt attempt_1475155149952_0002_m_000000_0 2016-09-29 21:25:01,311 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1475155149952_0002_m_000000_0 2016-09-29 21:25:01,311 INFO [ContainerLauncher #0] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : sq-hbase2.800best.com:3002 2016-09-29 21:25:01,350 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1475155149952_0002_m_000000_0 : 13562 2016-09-29 21:25:01,351 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1475155149952_0002_m_000000_0] using containerId: [container_1475155149952_0002_01_000002 on NM: [sq-hbase2.800best.com:3002] 2016-09-29 21:25:01,353 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_m_000000_0 TaskAttempt Transitioned from ASSIGNED to RUNNING 2016-09-29 21:25:01,354 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator: ATTEMPT_START task_1475155149952_0002_m_000000 2016-09-29 21:25:01,354 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1475155149952_0002_m_000000 Task Transitioned from SCHEDULED to RUNNING 2016-09-29 21:25:02,190 INFO [Socket Reader #1 for port 26213] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1475155149952_0002 (auth:SIMPLE) 2016-09-29 21:25:02,202 INFO [IPC Server handler 0 on 26213] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1475155149952_0002_m_000002 asked for a task 2016-09-29 21:25:02,202 INFO [IPC Server handler 0 on 26213] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1475155149952_0002_m_000002 given task: attempt_1475155149952_0002_m_000000_0 2016-09-29 21:25:02,258 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1475155149952_0002: ask=4 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:189952, vCores:1> knownNMs=3 2016-09-29 21:25:02,856 INFO [IPC Server handler 1 on 26213] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1475155149952_0002_m_000000_0 is : 0.0 2016-09-29 21:25:02,963 INFO [IPC Server handler 2 on 26213] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1475155149952_0002_m_000000_0 is : 1.0 2016-09-29 21:25:02,968 INFO [IPC Server handler 4 on 26213] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Done acknowledgement from attempt_1475155149952_0002_m_000000_0 2016-09-29 21:25:02,969 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_m_000000_0 TaskAttempt Transitioned from RUNNING to SUCCESS_CONTAINER_CLEANUP 2016-09-29 21:25:02,970 INFO [ContainerLauncher #1] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1475155149952_0002_01_000002 taskAttempt attempt_1475155149952_0002_m_000000_0 2016-09-29 21:25:02,970 INFO [ContainerLauncher #1] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1475155149952_0002_m_000000_0 2016-09-29 21:25:02,970 INFO [ContainerLauncher #1] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : sq-hbase2.800best.com:3002 2016-09-29 21:25:02,990 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_m_000000_0 TaskAttempt Transitioned from SUCCESS_CONTAINER_CLEANUP to SUCCEEDED 2016-09-29 21:25:02,995 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: Task succeeded with attempt attempt_1475155149952_0002_m_000000_0 2016-09-29 21:25:02,995 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1475155149952_0002_m_000000 Task Transitioned from RUNNING to SUCCEEDED 2016-09-29 21:25:02,997 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Num completed Tasks: 1 2016-09-29 21:25:03,258 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:1 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:03,261 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:189952, vCores:1> 2016-09-29 21:25:03,261 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold reached. Scheduling reduces. 2016-09-29 21:25:03,261 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: All maps assigned. Ramping up all remaining reduces:1 2016-09-29 21:25:03,261 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:1 AssignedMaps:1 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:1 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:04,266 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1475155149952_0002: ask=1 release= 0 newContainers=0 finishedContainers=1 resourcelimit=<memory:217088, vCores:1> knownNMs=3 2016-09-29 21:25:04,266 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1475155149952_0002_01_000002 2016-09-29 21:25:04,267 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:1 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:1 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:04,267 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1475155149952_0002_m_000000_0: Container killed by the ApplicationMaster. Container killed on request. Exit code is 143 Container exited with a non-zero exit code 143 2016-09-29 21:25:05,273 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 1 2016-09-29 21:25:05,273 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned to reduce 2016-09-29 21:25:05,273 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1475155149952_0002_01_000003 to attempt_1475155149952_0002_r_000000_0 2016-09-29 21:25:05,273 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:1 CompletedMaps:1 CompletedReds:0 ContAlloc:2 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:05,277 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved sq-hbase3.800best.com to /default-rack 2016-09-29 21:25:05,277 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_0 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED 2016-09-29 21:25:05,278 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1475155149952_0002_01_000003 taskAttempt attempt_1475155149952_0002_r_000000_0 2016-09-29 21:25:05,278 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1475155149952_0002_r_000000_0 2016-09-29 21:25:05,278 INFO [ContainerLauncher #2] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : sq-hbase3.800best.com:63020 2016-09-29 21:25:05,289 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1475155149952_0002_r_000000_0 : 13562 2016-09-29 21:25:05,289 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1475155149952_0002_r_000000_0] using containerId: [container_1475155149952_0002_01_000003 on NM: [sq-hbase3.800best.com:63020] 2016-09-29 21:25:05,289 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_0 TaskAttempt Transitioned from ASSIGNED to RUNNING 2016-09-29 21:25:05,290 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator: ATTEMPT_START task_1475155149952_0002_r_000000 2016-09-29 21:25:05,290 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1475155149952_0002_r_000000 Task Transitioned from SCHEDULED to RUNNING 2016-09-29 21:25:06,275 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1475155149952_0002: ask=1 release= 0 newContainers=0 finishedContainers=1 resourcelimit=<memory:217088, vCores:1> knownNMs=3 2016-09-29 21:25:06,275 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1475155149952_0002_01_000003 2016-09-29 21:25:06,276 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:2 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:06,276 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_0 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP 2016-09-29 21:25:06,276 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1475155149952_0002_r_000000_0: org.apache.commons.codec.binary.Base64.encodeBase64String([B)Ljava/lang/String; 2016-09-29 21:25:06,276 INFO [ContainerLauncher #3] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1475155149952_0002_01_000003 taskAttempt attempt_1475155149952_0002_r_000000_0 2016-09-29 21:25:06,277 INFO [ContainerLauncher #3] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1475155149952_0002_r_000000_0 2016-09-29 21:25:06,277 INFO [ContainerLauncher #3] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : sq-hbase3.800best.com:63020 2016-09-29 21:25:06,293 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_0 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP 2016-09-29 21:25:06,293 INFO [CommitterEvent Processor #1] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT 2016-09-29 21:25:06,297 WARN [CommitterEvent Processor #1] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://10.8.12.17:9000/linecount/win_output/_temporary/1/_temporary/attempt_1475155149952_0002_r_000000_0 2016-09-29 21:25:06,298 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_0 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED 2016-09-29 21:25:06,301 INFO [Thread-51] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures on node sq-hbase3.800best.com 2016-09-29 21:25:06,301 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_1 TaskAttempt Transitioned from NEW to UNASSIGNED 2016-09-29 21:25:07,276 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:2 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:07,277 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:217088, vCores:1> 2016-09-29 21:25:07,277 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: All maps assigned. Ramping up all remaining reduces:1 2016-09-29 21:25:07,277 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:1 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:2 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:08,280 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1475155149952_0002: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:217088, vCores:1> knownNMs=3 2016-09-29 21:25:09,283 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 1 2016-09-29 21:25:09,283 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned to reduce 2016-09-29 21:25:09,284 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1475155149952_0002_01_000004 to attempt_1475155149952_0002_r_000000_1 2016-09-29 21:25:09,284 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:1 CompletedMaps:1 CompletedReds:0 ContAlloc:3 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:09,284 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved sq-hbase3.800best.com to /default-rack 2016-09-29 21:25:09,284 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_1 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED 2016-09-29 21:25:09,285 INFO [ContainerLauncher #4] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1475155149952_0002_01_000004 taskAttempt attempt_1475155149952_0002_r_000000_1 2016-09-29 21:25:09,285 INFO [ContainerLauncher #4] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1475155149952_0002_r_000000_1 2016-09-29 21:25:09,285 INFO [ContainerLauncher #4] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : sq-hbase3.800best.com:63020 2016-09-29 21:25:09,293 INFO [ContainerLauncher #4] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1475155149952_0002_r_000000_1 : 13562 2016-09-29 21:25:09,293 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1475155149952_0002_r_000000_1] using containerId: [container_1475155149952_0002_01_000004 on NM: [sq-hbase3.800best.com:63020] 2016-09-29 21:25:09,293 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_1 TaskAttempt Transitioned from ASSIGNED to RUNNING 2016-09-29 21:25:09,293 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator: ATTEMPT_START task_1475155149952_0002_r_000000 2016-09-29 21:25:10,285 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1475155149952_0002: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:189952, vCores:1> knownNMs=3 2016-09-29 21:25:11,287 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1475155149952_0002_01_000004 2016-09-29 21:25:11,287 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:3 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:11,287 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_1 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP 2016-09-29 21:25:11,287 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1475155149952_0002_r_000000_1: org.apache.commons.codec.binary.Base64.encodeBase64String([B)Ljava/lang/String; 2016-09-29 21:25:11,288 INFO [ContainerLauncher #5] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1475155149952_0002_01_000004 taskAttempt attempt_1475155149952_0002_r_000000_1 2016-09-29 21:25:11,288 INFO [ContainerLauncher #5] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1475155149952_0002_r_000000_1 2016-09-29 21:25:11,288 INFO [ContainerLauncher #5] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : sq-hbase3.800best.com:63020 2016-09-29 21:25:11,296 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_1 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP 2016-09-29 21:25:11,296 INFO [CommitterEvent Processor #2] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT 2016-09-29 21:25:11,298 WARN [CommitterEvent Processor #2] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://10.8.12.17:9000/linecount/win_output/_temporary/1/_temporary/attempt_1475155149952_0002_r_000000_1 2016-09-29 21:25:11,298 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_1 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED 2016-09-29 21:25:11,299 INFO [Thread-51] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 2 failures on node sq-hbase3.800best.com 2016-09-29 21:25:11,299 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_2 TaskAttempt Transitioned from NEW to UNASSIGNED 2016-09-29 21:25:12,287 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:3 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:12,289 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:217088, vCores:1> 2016-09-29 21:25:12,289 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: All maps assigned. Ramping up all remaining reduces:1 2016-09-29 21:25:12,289 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:1 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:3 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:13,290 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1475155149952_0002: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:217088, vCores:1> knownNMs=3 2016-09-29 21:25:14,294 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 1 2016-09-29 21:25:14,294 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned to reduce 2016-09-29 21:25:14,294 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1475155149952_0002_01_000005 to attempt_1475155149952_0002_r_000000_2 2016-09-29 21:25:14,294 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:1 CompletedMaps:1 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:14,294 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved sq-hbase3.800best.com to /default-rack 2016-09-29 21:25:14,294 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_2 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED 2016-09-29 21:25:14,295 INFO [ContainerLauncher #6] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1475155149952_0002_01_000005 taskAttempt attempt_1475155149952_0002_r_000000_2 2016-09-29 21:25:14,295 INFO [ContainerLauncher #6] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1475155149952_0002_r_000000_2 2016-09-29 21:25:14,295 INFO [ContainerLauncher #6] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : sq-hbase3.800best.com:63020 2016-09-29 21:25:14,305 INFO [ContainerLauncher #6] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1475155149952_0002_r_000000_2 : 13562 2016-09-29 21:25:14,305 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1475155149952_0002_r_000000_2] using containerId: [container_1475155149952_0002_01_000005 on NM: [sq-hbase3.800best.com:63020] 2016-09-29 21:25:14,305 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_2 TaskAttempt Transitioned from ASSIGNED to RUNNING 2016-09-29 21:25:14,305 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator: ATTEMPT_START task_1475155149952_0002_r_000000 2016-09-29 21:25:15,296 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1475155149952_0002: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:189952, vCores:1> knownNMs=3 2016-09-29 21:25:16,298 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1475155149952_0002_01_000005 2016-09-29 21:25:16,298 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:16,298 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_2 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP 2016-09-29 21:25:16,298 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1475155149952_0002_r_000000_2: org.apache.commons.codec.binary.Base64.encodeBase64String([B)Ljava/lang/String; 2016-09-29 21:25:16,299 INFO [ContainerLauncher #7] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1475155149952_0002_01_000005 taskAttempt attempt_1475155149952_0002_r_000000_2 2016-09-29 21:25:16,299 INFO [ContainerLauncher #7] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1475155149952_0002_r_000000_2 2016-09-29 21:25:16,299 INFO [ContainerLauncher #7] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : sq-hbase3.800best.com:63020 2016-09-29 21:25:16,305 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_2 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP 2016-09-29 21:25:16,306 INFO [CommitterEvent Processor #3] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT 2016-09-29 21:25:16,307 WARN [CommitterEvent Processor #3] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://10.8.12.17:9000/linecount/win_output/_temporary/1/_temporary/attempt_1475155149952_0002_r_000000_2 2016-09-29 21:25:16,307 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_2 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED 2016-09-29 21:25:16,308 INFO [Thread-51] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 3 failures on node sq-hbase3.800best.com 2016-09-29 21:25:16,308 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_3 TaskAttempt Transitioned from NEW to UNASSIGNED 2016-09-29 21:25:16,308 INFO [Thread-51] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Blacklisted host sq-hbase3.800best.com 2016-09-29 21:25:17,298 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:17,300 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Update the blacklist for application_1475155149952_0002: blacklistAdditions=1 blacklistRemovals=0 2016-09-29 21:25:17,300 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Ignore blacklisting set to true. Known: 3, Blacklisted: 1, 33% 2016-09-29 21:25:17,301 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:217088, vCores:1> 2016-09-29 21:25:17,301 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: All maps assigned. Ramping up all remaining reduces:1 2016-09-29 21:25:17,301 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:1 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:18,302 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1475155149952_0002: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:217088, vCores:1> knownNMs=3 2016-09-29 21:25:18,302 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Update the blacklist for application_1475155149952_0002: blacklistAdditions=0 blacklistRemovals=1 2016-09-29 21:25:19,311 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 1 2016-09-29 21:25:19,311 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned to reduce 2016-09-29 21:25:19,311 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1475155149952_0002_01_000006 to attempt_1475155149952_0002_r_000000_3 2016-09-29 21:25:19,311 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:1 CompletedMaps:1 CompletedReds:0 ContAlloc:5 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:19,311 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved sq-hbase3.800best.com to /default-rack 2016-09-29 21:25:19,312 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_3 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED 2016-09-29 21:25:19,313 INFO [ContainerLauncher #8] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1475155149952_0002_01_000006 taskAttempt attempt_1475155149952_0002_r_000000_3 2016-09-29 21:25:19,313 INFO [ContainerLauncher #8] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1475155149952_0002_r_000000_3 2016-09-29 21:25:19,313 INFO [ContainerLauncher #8] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : sq-hbase3.800best.com:63020 2016-09-29 21:25:19,321 INFO [ContainerLauncher #8] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1475155149952_0002_r_000000_3 : 13562 2016-09-29 21:25:19,321 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1475155149952_0002_r_000000_3] using containerId: [container_1475155149952_0002_01_000006 on NM: [sq-hbase3.800best.com:63020] 2016-09-29 21:25:19,321 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_3 TaskAttempt Transitioned from ASSIGNED to RUNNING 2016-09-29 21:25:19,321 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator: ATTEMPT_START task_1475155149952_0002_r_000000 2016-09-29 21:25:20,313 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1475155149952_0002: ask=1 release= 0 newContainers=0 finishedContainers=1 resourcelimit=<memory:217088, vCores:1> knownNMs=3 2016-09-29 21:25:20,313 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1475155149952_0002_01_000006 2016-09-29 21:25:20,313 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:5 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:20,313 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_3 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP 2016-09-29 21:25:20,313 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1475155149952_0002_r_000000_3: org.apache.commons.codec.binary.Base64.encodeBase64String([B)Ljava/lang/String; 2016-09-29 21:25:20,313 INFO [ContainerLauncher #9] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1475155149952_0002_01_000006 taskAttempt attempt_1475155149952_0002_r_000000_3 2016-09-29 21:25:20,314 INFO [ContainerLauncher #9] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1475155149952_0002_r_000000_3 2016-09-29 21:25:20,314 INFO [ContainerLauncher #9] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : sq-hbase3.800best.com:63020 2016-09-29 21:25:20,320 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_3 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP 2016-09-29 21:25:20,320 INFO [CommitterEvent Processor #4] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT 2016-09-29 21:25:20,322 WARN [CommitterEvent Processor #4] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://10.8.12.17:9000/linecount/win_output/_temporary/1/_temporary/attempt_1475155149952_0002_r_000000_3 2016-09-29 21:25:20,322 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1475155149952_0002_r_000000_3 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED 2016-09-29 21:25:20,324 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1475155149952_0002_r_000000 Task Transitioned from RUNNING to FAILED 2016-09-29 21:25:20,324 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Num completed Tasks: 2 2016-09-29 21:25:20,324 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Job failed as tasks failed. failedMaps:0 failedReduces:1 2016-09-29 21:25:20,324 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1475155149952_0002Job Transitioned from RUNNING to FAIL_ABORT 2016-09-29 21:25:20,324 INFO [CommitterEvent Processor #0] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: JOB_ABORT 2016-09-29 21:25:20,331 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1475155149952_0002Job Transitioned from FAIL_ABORT to FAILED 2016-09-29 21:25:20,332 INFO [Thread-71] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: We are finishing cleanly so this is the last retry 2016-09-29 21:25:20,332 INFO [Thread-71] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify RMCommunicator isAMLastRetry: true 2016-09-29 21:25:20,332 INFO [Thread-71] org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator: RMCommunicator notified that shouldUnregistered is: true 2016-09-29 21:25:20,332 INFO [Thread-71] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify JHEH isAMLastRetry: true 2016-09-29 21:25:20,332 INFO [Thread-71] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: JobHistoryEventHandler notified that forceJobCompletion is true 2016-09-29 21:25:20,332 INFO [Thread-71] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Calling stop for all the services 2016-09-29 21:25:20,333 INFO [Thread-71] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Stopping JobHistoryEventHandler. Size of the outstanding queue size is 0 2016-09-29 21:25:20,360 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copying hdfs://ns:8020/tmp/hadoop-yarn/staging/appadmin/.staging/job_1475155149952_0002/job_1475155149952_0002_1.jhist to hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate/appadmin/job_1475155149952_0002-1475155493770-appadmin-linecount-1475155520324-1-0-FAILED-default-1475155499222.jhist_tmp 2016-09-29 21:25:20,382 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copied to done location: hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate/appadmin/job_1475155149952_0002-1475155493770-appadmin-linecount-1475155520324-1-0-FAILED-default-1475155499222.jhist_tmp 2016-09-29 21:25:20,384 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copying hdfs://ns:8020/tmp/hadoop-yarn/staging/appadmin/.staging/job_1475155149952_0002/job_1475155149952_0002_1_conf.xml to hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate/appadmin/job_1475155149952_0002_conf.xml_tmp 2016-09-29 21:25:20,407 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copied to done location: hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate/appadmin/job_1475155149952_0002_conf.xml_tmp 2016-09-29 21:25:20,418 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp to done: hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate/appadmin/job_1475155149952_0002.summary_tmp to hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate/appadmin/job_1475155149952_0002.summary 2016-09-29 21:25:20,420 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp to done: hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate/appadmin/job_1475155149952_0002_conf.xml_tmp to hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate/appadmin/job_1475155149952_0002_conf.xml 2016-09-29 21:25:20,422 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp to done: hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate/appadmin/job_1475155149952_0002-1475155493770-appadmin-linecount-1475155520324-1-0-FAILED-default-1475155499222.jhist_tmp to hdfs://ns:8020/tmp/hadoop-yarn/staging/history/done_intermediate/appadmin/job_1475155149952_0002-1475155493770-appadmin-linecount-1475155520324-1-0-FAILED-default-1475155499222.jhist 2016-09-29 21:25:20,422 INFO [Thread-71] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Stopped JobHistoryEventHandler. super.stop() 2016-09-29 21:25:20,424 INFO [Thread-71] org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator: Setting job diagnostics to Task failed task_1475155149952_0002_r_000000 Job failed as tasks failed. failedMaps:0 failedReduces:1 2016-09-29 21:25:20,425 INFO [Thread-71] org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator: History url is http://sq-hbase1.800best.com:19888/jobhistory/job/job_1475155149952_0002 2016-09-29 21:25:20,434 INFO [Thread-71] org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator: Waiting for application to be successfully unregistered. 2016-09-29 21:25:21,436 INFO [Thread-71] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Final Stats: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:5 ContRel:0 HostLocal:1 RackLocal:0 2016-09-29 21:25:21,436 INFO [Thread-71] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Deleting staging directory hdfs://10.8.12.17:9000 /tmp/hadoop-yarn/staging/appadmin/.staging/job_1475155149952_0002 2016-09-29 21:25:21,439 INFO [Thread-71] org.apache.hadoop.ipc.Server: Stopping server on 26213 2016-09-29 21:25:21,440 INFO [IPC Server listener on 26213] org.apache.hadoop.ipc.Server: Stopping IPC Server listener on 26213 2016-09-29 21:25:21,440 INFO [TaskHeartbeatHandler PingChecker] org.apache.hadoop.mapreduce.v2.app.TaskHeartbeatHandler: TaskHeartbeatHandler thread interrupted 2016-09-29 21:25:21,441 INFO [IPC Server Responder] org.apache.hadoop.ipc.Server: Stopping IPC Server Responder


Kaiming Wan

unread,
Sep 29, 2016, 9:29:25 AM9/29/16
to Alluxio Users

Kaiming Wan

unread,
Sep 29, 2016, 9:30:22 AM9/29/16
to Alluxio Users
Can anyone help me? Thanks!

Kaiming Wan

unread,
Sep 29, 2016, 10:48:20 PM9/29/16
to Alluxio Users
When I check the log file "yarn-appadmin-nodemanager-bs.best.com.log", it has error like:

2016-09-30 10:26:02,579 WARN org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch: Failed to launch container.
java
.lang.NoSuchMethodError: org.apache.commons.codec.binary.Base64.encodeBase64String([B)Ljava/lang/String;
 at org
.apache.hadoop.yarn.util.AuxiliaryServiceHelper.setServiceDataIntoEnv(AuxiliaryServiceHelper.java:45)
 at org
.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.sanitizeEnv(ContainerLaunch.java:823)
 at org
.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:268)
 at org
.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82)
 at java
.util.concurrent.FutureTask.run(FutureTask.java:266)
 at java
.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
 at java
.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
 at java
.lang.Thread.run(Thread.java:745)



Kaiming Wan

unread,
Sep 29, 2016, 11:02:29 PM9/29/16
to Alluxio Users
I use the following code to trace the "noSuchMethodError" in my MR job:

System.out.println(" codec location "+org.apache.commons.codec.binary.Base64.class.getProtectionDomain().getCodeSource());


it prints out :
 codec location (file:/home/appadmin/hadoop-2.7.2/share/hadoop/common/lib/alluxio-core-client-1.2.0-jar-with-dependencies.jar <no signer certificates>)

It is obvious that the Base64 class is in alluxio-core-client-1.2.0-jar-with-dependencies.jar.

Does it mean may be there is something wrong when I compile alluxio.

Kaiming Wan

unread,
Sep 30, 2016, 1:12:52 AM9/30/16
to Alluxio Users
I have solved this issue. The problem is caused by dependency "commons-codec".

I think this is a bug. If I compile alluxio without changing the commons-codec version larger than 1.3, when I run MR job it will gives out the error:"java.lang.NoSuchMethodError: org.apache.commons.codec.binary.Base64.encodeBase64String([B)Ljava/lang/String;"

This is because commons-codec jar smaller than version 1.4 does not have the method encodeBase64String in class Base64.

I change all the commons-codec dependency and set the version to 1.9.  

Following files are changed:
  1. $ALLUXIO_HOME/pom.xml
  2. $ALLUXIO_HOME/core/server/pom.xml
  3. $ALLUXIO_HOME/underfs/s3/pom.xml
  4. $ALLUXIO_HOME/underfs/gcs/pom.xml

Bin Fan

unread,
Sep 30, 2016, 2:35:51 AM9/30/16
to Kaiming Wan, Alluxio Users
Thanks Kaiming for reporting the issue and providing the solution too!

I am working on this and will update you here for any further patch on this

--
You received this message because you are subscribed to the Google Groups "Alluxio Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to alluxio-users+unsubscribe@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Bin Fan

unread,
Sep 30, 2016, 8:35:52 PM9/30/16
to Alluxio Users, wan...@gmail.com
I merged a fix to 1.3-release candidate. This should apply to the same code in 1.2.

https://github.com/Alluxio/alluxio/pull/3985

Could you take a look and give a try? thanks

- Bin
Reply all
Reply to author
Forward
0 new messages