2014-09-25 09:51:15,905 INFO [main] org.apache.hadoop.streaming.PipeMapRed: PipeMapRed failed! java.lang.RuntimeException: java.io.EOFException at org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:344) at org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:543) at org.apache.hadoop.streaming.PipeReducer.close(PipeReducer.java:134) at org.apache.hadoop.mapred.Task$OldCombinerRunner.combine(Task.java:1577) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpill(MapTask.java:1631) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:1482) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:440) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:167) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) Caused by: java.io.EOFException at java.io.DataInputStream.readFully(DataInputStream.java:197) at org.apache.hadoop.typedbytes.TypedBytesInput.readRawBytes(TypedBytesInput.java:218) at org.apache.hadoop.typedbytes.TypedBytesInput.readRaw(TypedBytesInput.java:152) at org.apache.hadoop.typedbytes.TypedBytesInput.readRawVector(TypedBytesInput.java:412) at org.apache.hadoop.typedbytes.TypedBytesInput.readRaw(TypedBytesInput.java:144) at org.apache.hadoop.streaming.io.TypedBytesOutputReader.readKeyValue(TypedBytesOutputReader.java:56) at org.apache.hadoop.streaming.PipeMapRed$MROutputThread.run(PipeMapRed.java:386) 2014-09-25 09:51:15,910 WARN [main] org.apache.hadoop.mapred.YarnChild: Exception running child : java.lang.RuntimeException: java.io.EOFException at org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:344) at org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:543) at org.apache.hadoop.streaming.PipeReducer.close(PipeReducer.java:134) at org.apache.hadoop.mapred.Task$OldCombinerRunner.combine(Task.java:1577) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpill(MapTask.java:1631) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:1482) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:440) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:167) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) Caused by: java.io.EOFException at java.io.DataInputStream.readFully(DataInputStream.java:197) at org.apache.hadoop.typedbytes.TypedBytesInput.readRawBytes(TypedBytesInput.java:218) at org.apache.hadoop.typedbytes.TypedBytesInput.readRaw(TypedBytesInput.java:152) at org.apache.hadoop.typedbytes.TypedBytesInput.readRawVector(TypedBytesInput.java:412) at org.apache.hadoop.typedbytes.TypedBytesInput.readRaw(TypedBytesInput.java:144) at org.apache.hadoop.streaming.io.TypedBytesOutputReader.readKeyValue(TypedBytesOutputReader.java:56) at org.apache.hadoop.streaming.PipeMapRed$MROutputThread.run(PipeMapRed.java:386)
Don't really know how to proceed, is there a way how I could debug count.cols function?
Best,
Kirsti
--
post: rha...@googlegroups.com ||
unsubscribe: rhadoop+u...@googlegroups.com ||
web: https://groups.google.com/d/forum/rhadoop?hl=en-US
---
You received this message because you are subscribed to the Google Groups "RHadoop" group.
To unsubscribe from this group and stop receiving emails from it, send an email to rhadoop+u...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
Hi,
aa is data.frame. I tested a=as.data.frame(count.cols(input(aa))) and sometimes it works, sometime it fails (with the same data.frame). I was using 0.3.0, but tried this with 0.4.0.
Apparently in 0.4.0 there is not anymore count.cols but just countn there I couldn't get
xx=to.dfs(aa)
a=as.data.frame(plyrmr::count(input(xx))) to work at all, complains that "no applicable method for 'as.pipermr' applied to an object of class "function".
Additionally, without converting to dfs
a=as.data.frame(plyrmr::count(input(aa))), I got some memory problems and when tested, I couldn't run any mapreduce there as I got error:
MAP capability required is more than the supported max container capability in the cluster. Killing the Job. mapResourceReqt: 4096 maxContainerCapability:3072
Job received Kill while in RUNNING state.
REDUCE capability required is more than the supported max container capability in the cluster. Killing the Job. reduceResourceReqt: 4096 maxContainerCapability:3072
Anything I could do? How to use objects in dfs with count function in with 0.4.0?