'c_id', 'bp', 'rr', 'v_id', 'm_id', 'id', 'timestamp']]"]["/user/root/raw_data/"]"] 2012-09-07 05:03:12,514 INFO cascading.flow.hadoop.FlowMapper: sinking to: GroupBy(success_bid)[by:[{1}:'id']] 2012-09-07 05:03:12,590 ERROR cascading.flow.stream.TrapHandler: caught Throwable, no trap available, rethrowing cascading.tuple.TupleException: unable to read from input identifier: hdfs:/url/user/root/raw_data/data201.gz at cascading.tuple.TupleEntrySchemeIterator.hasNext(TupleEntrySchemeIterator.java:127) at cascading.flow.stream.SourceStage.map(SourceStage.java:76) at cascading.flow.stream.SourceStage.run(SourceStage.java:58) at cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:127) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:393) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:327) at org.apache.hadoop.mapred.Child$4.run(Child.java:270) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) at org.apache.hadoop.mapred.Child.main(Child.java:264) Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:473) at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:565) at java.io.DataInputStream.read(DataInputStream.java:132) at org.apache.hadoop.io.compress.DecompressorStream.getCompressedData(DecompressorStream.java:157) at org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:102) at org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:83) at java.io.InputStream.read(InputStream.java:82) at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:209) at org.apache.hadoop.util.LineReader.readLine(LineReader.java:173) at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:160) at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:38) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.moveToNext(MapTask.java:210) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.next(MapTask.java:195) at cascading.tap.hadoop.util.MeasuredRecordReader.next(MeasuredRecordReader.java:61) at cascading.scheme.hadoop.TextDelimited.source(TextDelimited.java:757) at cascading.tuple.TupleEntrySchemeIterator.getNext(TupleEntrySchemeIterator.java:140) at cascading.tuple.TupleEntrySchemeIterator.hasNext(TupleEntrySchemeIterator.java:120) ... 10 more 2012-09-07 05:03:12,596 ERROR cascading.flow.stream.SourceStage: caught throwable cascading.tuple.TupleException: unable to read from input identifier: hdfs:/url/user/root/raw_data/data201.gz
at cascading.tuple.TupleEntrySchemeIterator.hasNext(TupleEntrySchemeIterator.java:127) at cascading.flow.stream.SourceStage.map(SourceStage.java:76) at cascading.flow.stream.SourceStage.run(SourceStage.java:58) at cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:127) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:393) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:327) at org.apache.hadoop.mapred.Child$4.run(Child.java:270) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) at org.apache.hadoop.mapred.Child.main(Child.java:264) Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:473) at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:565) at java.io.DataInputStream.read(DataInputStream.java:132) at org.apache.hadoop.io.compress.DecompressorStream.getCompressedData(DecompressorStream.java:157) at org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:102) at org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:83) at java.io.InputStream.read(InputStream.java:82) at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:209) at org.apache.hadoop.util.LineReader.readLine(LineReader.java:173) at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:160) at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:38) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.moveToNext(MapTask.java:210) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.next(MapTask.java:195) at cascading.tap.hadoop.util.MeasuredRecordReader.next(MeasuredRecordReader.java:61) at cascading.scheme.hadoop.TextDelimited.source(TextDelimited.java:757) at cascading.tuple.TupleEntrySchemeIterator.getNext(TupleEntrySchemeIterator.java:140) at cascading.tuple.TupleEntrySchemeIterator.hasNext(TupleEntrySchemeIterator.java:120) ... 10 more 2012-09-07 05:03:12,597 WARN cascading.flow.stream.SourceStage: failed closing iterator java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:473) at org.apache.hadoop.hdfs.DFSInputStream.close(DFSInputStream.java:488) at java.io.FilterInputStream.close(FilterInputStream.java:155) at org.apache.hadoop.io.compress.DecompressorStream.close(DecompressorStream.java:199) at org.apache.hadoop.util.LineReader.close(LineReader.java:149) at org.apache.hadoop.mapred.LineRecordReader.close(LineRecordReader.java:195) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.close(MapTask.java:217) at cascading.tap.hadoop.io.RecordReaderIterator.close(RecordReaderIterator.java:45) at cascading.tuple.TupleEntrySchemeIterator.close(TupleEntrySchemeIterator.java:201) at cascading.flow.stream.SourceStage.map(SourceStage.java:117) at cascading.flow.stream.SourceStage.run(SourceStage.java:58) at cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:127) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:393) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:327) at org.apache.hadoop.mapred.Child$4.run(Child.java:270) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) at org.apache.hadoop.mapred.Child.main(Child.java:264) 2012-09-07 05:03:12,611 INFO org.apache.hadoop.mapred.TaskLogsTruncater: Initializing logs' truncater with mapRetainSize=-1 and reduceRetainSize=-1 2012-09-07 05:03:12,614 ERROR org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:ads (auth:SIMPLE) cause:java.io.IOException: Filesystem closed 2012-09-07 05:03:12,614 WARN org.apache.hadoop.mapred.Child: Error running child java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:473) at org.apache.hadoop.hdfs.DFSInputStream.close(DFSInputStream.java:488) at java.io.FilterInputStream.close(FilterInputStream.java:155) at org.apache.hadoop.io.compress.DecompressorStream.close(DecompressorStream.java:199) at org.apache.hadoop.util.LineReader.close(LineReader.java:149) at org.apache.hadoop.mapred.LineRecordReader.close(LineRecordReader.java:195) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.close(MapTask.java:217) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:397) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:327) at org.apache.hadoop.mapred.Child$4.run(Child.java:270) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) at org.apache.hadoop.mapred.Child.main(Child.java:264) 2012-09-07 05:03:12,627 WARN org.apache.hadoop.mapred.Task: Parent died. Exiting attempt_201209051416_0136_m_002942_0