empty data file exception

255 views
Skip to first unread message

Ben

unread,
Sep 7, 2012, 2:21:03 PM9/7/12
to cascadi...@googlegroups.com
I have a source data path and have one empty file (data201.gz) and a bunch of other nonempty files under the path (/user/root/raw_data). When reading the data from the path, I got the following exception. How can I avoid this exception? 

Thanks! 

Ben

'c_id', 'bp', 'rr', 'v_id', 'm_id', 'id', 'timestamp']]"]["/user/root/raw_data/"]"]
2012-09-07 05:03:12,514 INFO cascading.flow.hadoop.FlowMapper: sinking to: GroupBy(success_bid)[by:[{1}:'id']]
2012-09-07 05:03:12,590 ERROR cascading.flow.stream.TrapHandler: caught Throwable, no trap available, rethrowing
cascading.tuple.TupleException: unable to read from input identifier: hdfs:/url/user/root/raw_data/data201.gz
	at cascading.tuple.TupleEntrySchemeIterator.hasNext(TupleEntrySchemeIterator.java:127)
	at cascading.flow.stream.SourceStage.map(SourceStage.java:76)
	at cascading.flow.stream.SourceStage.run(SourceStage.java:58)
	at cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:127)
	at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:393)
	at org.apache.hadoop.mapred.MapTask.run(MapTask.java:327)
	at org.apache.hadoop.mapred.Child$4.run(Child.java:270)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:396)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232)
	at org.apache.hadoop.mapred.Child.main(Child.java:264)
Caused by: java.io.IOException: Filesystem closed
	at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:473)
	at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:565)
	at java.io.DataInputStream.read(DataInputStream.java:132)
	at org.apache.hadoop.io.compress.DecompressorStream.getCompressedData(DecompressorStream.java:157)
	at org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:102)
	at org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:83)
	at java.io.InputStream.read(InputStream.java:82)
	at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:209)
	at org.apache.hadoop.util.LineReader.readLine(LineReader.java:173)
	at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:160)
	at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:38)
	at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.moveToNext(MapTask.java:210)
	at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.next(MapTask.java:195)
	at cascading.tap.hadoop.util.MeasuredRecordReader.next(MeasuredRecordReader.java:61)
	at cascading.scheme.hadoop.TextDelimited.source(TextDelimited.java:757)
	at cascading.tuple.TupleEntrySchemeIterator.getNext(TupleEntrySchemeIterator.java:140)
	at cascading.tuple.TupleEntrySchemeIterator.hasNext(TupleEntrySchemeIterator.java:120)
	... 10 more
2012-09-07 05:03:12,596 ERROR cascading.flow.stream.SourceStage: caught throwable
cascading.tuple.TupleException: unable to read from input identifier: hdfs:/url/user/root/raw_data/data201.gz
at cascading.tuple.TupleEntrySchemeIterator.hasNext(TupleEntrySchemeIterator.java:127) at cascading.flow.stream.SourceStage.map(SourceStage.java:76) at cascading.flow.stream.SourceStage.run(SourceStage.java:58) at cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:127) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:393) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:327) at org.apache.hadoop.mapred.Child$4.run(Child.java:270) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) at org.apache.hadoop.mapred.Child.main(Child.java:264) Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:473) at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:565) at java.io.DataInputStream.read(DataInputStream.java:132) at org.apache.hadoop.io.compress.DecompressorStream.getCompressedData(DecompressorStream.java:157) at org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:102) at org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:83) at java.io.InputStream.read(InputStream.java:82) at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:209) at org.apache.hadoop.util.LineReader.readLine(LineReader.java:173) at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:160) at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:38) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.moveToNext(MapTask.java:210) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.next(MapTask.java:195) at cascading.tap.hadoop.util.MeasuredRecordReader.next(MeasuredRecordReader.java:61) at cascading.scheme.hadoop.TextDelimited.source(TextDelimited.java:757) at cascading.tuple.TupleEntrySchemeIterator.getNext(TupleEntrySchemeIterator.java:140) at cascading.tuple.TupleEntrySchemeIterator.hasNext(TupleEntrySchemeIterator.java:120) ... 10 more 2012-09-07 05:03:12,597 WARN cascading.flow.stream.SourceStage: failed closing iterator java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:473) at org.apache.hadoop.hdfs.DFSInputStream.close(DFSInputStream.java:488) at java.io.FilterInputStream.close(FilterInputStream.java:155) at org.apache.hadoop.io.compress.DecompressorStream.close(DecompressorStream.java:199) at org.apache.hadoop.util.LineReader.close(LineReader.java:149) at org.apache.hadoop.mapred.LineRecordReader.close(LineRecordReader.java:195) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.close(MapTask.java:217) at cascading.tap.hadoop.io.RecordReaderIterator.close(RecordReaderIterator.java:45) at cascading.tuple.TupleEntrySchemeIterator.close(TupleEntrySchemeIterator.java:201) at cascading.flow.stream.SourceStage.map(SourceStage.java:117) at cascading.flow.stream.SourceStage.run(SourceStage.java:58) at cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:127) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:393) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:327) at org.apache.hadoop.mapred.Child$4.run(Child.java:270) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) at org.apache.hadoop.mapred.Child.main(Child.java:264) 2012-09-07 05:03:12,611 INFO org.apache.hadoop.mapred.TaskLogsTruncater: Initializing logs' truncater with mapRetainSize=-1 and reduceRetainSize=-1 2012-09-07 05:03:12,614 ERROR org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:ads (auth:SIMPLE) cause:java.io.IOException: Filesystem closed 2012-09-07 05:03:12,614 WARN org.apache.hadoop.mapred.Child: Error running child java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:473) at org.apache.hadoop.hdfs.DFSInputStream.close(DFSInputStream.java:488) at java.io.FilterInputStream.close(FilterInputStream.java:155) at org.apache.hadoop.io.compress.DecompressorStream.close(DecompressorStream.java:199) at org.apache.hadoop.util.LineReader.close(LineReader.java:149) at org.apache.hadoop.mapred.LineRecordReader.close(LineRecordReader.java:195) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.close(MapTask.java:217) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:397) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:327) at org.apache.hadoop.mapred.Child$4.run(Child.java:270) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) at org.apache.hadoop.mapred.Child.main(Child.java:264) 2012-09-07 05:03:12,627 WARN org.apache.hadoop.mapred.Task: Parent died. Exiting attempt_201209051416_0136_m_002942_0


Reply all
Reply to author
Forward
0 new messages