14/03/18 21:02:54 INFO mapreduce.Job: map 100% reduce 67% 14/03/18 22:24:32 INFO mapreduce.Job: Task Id : attempt_1395174999269_0003_r_000073_0, Status : FAILED Container [pid=20617,containerID=container_1395174999269_0003_01_001031] is running beyond physical memory limits. Current usage: 9.1 GB of 8 GB physical memory used; 9.8 GB of 16.8 GB virtual memory used. Killing container. Dump of the process-tree for container_1395174999269_0003_01_001031 : |- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL_CMD_LINE |- 20617 3473 20617 20617 (bash) 0 0 65425408 276 /bin/bash -c /usr/jdk64/jdk1.6.0_31/bin/java -Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN -Xmx2048m -Djava.io.tmpdir=/hadooptmp/yarn/local/usercache/harshav/appcache/application_1395174999269_0003/container_1395174999269_0003_01_001031/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/hadoop/yarn/log/application_1395174999269_0003/container_1395174999269_0003_01_001031 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.6.46 38160 attempt_1395174999269_0003_r_000073_0 1031 1>/hadoop/yarn/log/application_1395174999269_0003/container_1395174999269_0003_01_001031/stdout 2>/hadoop/yarn/log/application_1395174999269_0003/container_1395174999269_0003_01_001031/stderr |- 20628 20617 20617 20617 (java) 13496 815 2571829248 511248 /usr/jdk64/jdk1.6.0_31/bin/java -Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN -Xmx2048m -Djava.io.tmpdir=/hadooptmp/yarn/local/usercache/harshav/appcache/application_1395174999269_0003/container_1395174999269_0003_01_001031/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/hadoop/yarn/log/application_1395174999269_0003/container_1395174999269_0003_01_001031 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.6.46 38160 attempt_1395174999269_0003_r_000073_0 1031 |- 23513 23495 20617 20617 (cat) 29 185 60411904 133 cat |- 23495 20628 20617 20617 (R) 487542 6828 7784603648 1861984 /usr/local/lib64/R/bin/exec/R --slave --no-restore --file=./rmr-streaming-reduce77093ef269ef --args Container killed on request. Exit code is 143
--
post: rha...@googlegroups.com ||
unsubscribe: rhadoop+u...@googlegroups.com ||
web: https://groups.google.com/d/forum/rhadoop?hl=en-US
---
You received this message because you are subscribed to the Google Groups "RHadoop" group.
To unsubscribe from this group and stop receiving emails from it, send an email to rhadoop+u...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
Here is the tail of the output from the rmr.str(gc()) call from a reducer that failed due to memory error.
gc() num [1:2, 1:6] 7.38e+05 3.82e+06 3.95e+01 2.92e+01 1.27e+06 ... - attr(*, "dimnames")=List of 2 ..$ : chr [1:2] "Ncells" "Vcells" ..$ : chr [1:6] "used" "(Mb)" "gc trigger" "(Mb)" ... Dotted pair list of 13 $ : language (function() { invisible(if (is.null(formals(load)$verbose)) load("./rmr-local-env50aa18c54f41") else load("./rmr-local-env50aa18c54f41", ... $ : language rmr2:::reduce.loop(reduce = reduce, vectorized = vectorized.reduce, keyval.reader = default.reader(), keyval.writer = output.writer(), profile = profile.nodes) $ : language apply.reduce(complete, red.as.kv) $ : language c.keyval(reduce.keyval(kv, reduce)) $ : language reduce.keyval(kv, reduce) $ : language mapply(FUN, keys(kvs), values(kvs), SIMPLIFY = FALSE) $ : language (function (...) do.call(FUN, c(.orig, list(...))))(dots[[1L]][[14L]], dots[[2L]][[14L]]) $ : language do.call(FUN, c(.orig, list(...))) $ : language (function (k, vv, reduce) as.keyval(reduce(k, vv)))(reduce = function (k, v) ... $ : language as.keyval(reduce(k, vv)) $ : language is.keyval(x) $ : language reduce(k, vv) $ :length 2 rmr.str(gc()) ..- attr(*, "srcref")=Class 'srcref' atomic [1:8] 22 5 22 17 5 17 22 22 .. .. ..- attr(*, "srcfile")=Classes 'srcfilecopy', 'srcfile' <environment: 0x1e0c7440> gc() num [1:2, 1:6] 7.38e+05 3.82e+06 3.95e+01 2.92e+01 1.27e+06 ... - attr(*, "dimnames")=List of 2 ..$ : chr [1:2] "Ncells" "Vcells" ..$ : chr [1:6] "used" "(Mb)" "gc trigger" "(Mb)" ... Dotted pair list of 13 $ : language (function() { invisible(if (is.null(formals(load)$verbose)) load("./rmr-local-env50aa18c54f41") else load("./rmr-local-env50aa18c54f41", ... $ : language rmr2:::reduce.loop(reduce = reduce, vectorized = vectorized.reduce, keyval.reader = default.reader(), keyval.writer = output.writer(), profile = profile.nodes) $ : language apply.reduce(complete, red.as.kv) $ : language c.keyval(reduce.keyval(kv, reduce)) $ : language reduce.keyval(kv, reduce) $ : language mapply(FUN, keys(kvs), values(kvs), SIMPLIFY = FALSE) $ : language (function (...) do.call(FUN, c(.orig, list(...))))(dots[[1L]][[15L]], dots[[2L]][[15L]]) $ : language do.call(FUN, c(.orig, list(...))) $ : language (function (k, vv, reduce) as.keyval(reduce(k, vv)))(reduce = function (k, v) ... $ : language as.keyval(reduce(k, vv)) $ : language is.keyval(x) $ : language reduce(k, vv) $ :length 2 rmr.str(gc()) ..- attr(*, "srcref")=Class 'srcref' atomic [1:8] 22 5 22 17 5 17 22 22 .. .. ..- attr(*, "srcfile")=Classes 'srcfilecopy', 'srcfile' <environment: 0x1e0c7440> gc() num [1:2, 1:6] 7.38e+05 3.82e+06 3.95e+01 2.92e+01 1.27e+06 ... - attr(*, "dimnames")=List of 2 ..$ : chr [1:2] "Ncells" "Vcells" ..$ : chr [1:6] "used" "(Mb)" "gc trigger" "(Mb)" ...
I changed all the options to "mapreduce". However, I still get warnings saying "mapred.xxx.yyy is deprecated". Does it have to do with how rmr is calling hadoop streaming, i.e., using mapreduce version 1 APIs as opposed to the version 2?
14/03/19 19:46:15 INFO Configuration.deprecation: user.name is deprecated. Instead, use mapreduce.job.user.name 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.jar is deprecated. Instead, use mapreduce.job.jar 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.cache.files.filesizes is deprecated. Instead, use mapreduce.job.cache.files.filesizes 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.cache.files is deprecated. Instead, use mapreduce.job.cache.files 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.output.value.class is deprecated. Instead, use mapreduce.job.output.value.class 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.mapoutput.value.class is deprecated. Instead, use mapreduce.map.output.value.class 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.job.name is deprecated. Instead, use mapreduce.job.name 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.textoutputformat.separator is deprecated. Instead, use mapreduce.output.textoutputformat.separator 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.input.dir is deprecated. Instead, use mapreduce.input.fileinputformat.inputdir 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.output.dir is deprecated. Instead, use mapreduce.output.fileoutputformat.outputdir 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.map.tasks is deprecated. Instead, use mapreduce.job.maps 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.cache.files.timestamps is deprecated. Instead, use mapreduce.job.cache.files.timestamps 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.output.key.class is deprecated. Instead, use mapreduce.job.output.key.class 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.mapoutput.key.class is deprecated. Instead, use mapreduce.map.output.key.class 14/03/19 19:46:15 INFO Configuration.deprecation: mapred.working.dir is deprecated. Instead, use mapreduce.job.working.dir
--