2017-06-30T08:10:02,161 INFO [Thread-21] org.apache.hadoop.mapred.LocalJobRunner - reduce task executor complete.
2017-06-30T08:10:02,181 INFO [communication thread] org.apache.hadoop.mapred.LocalJobRunner - reduce > reduce
2017-06-30T08:10:02,186 INFO [communication thread] org.apache.hadoop.mapred.LocalJobRunner - reduce > reduce
2017-06-30T08:10:02,170 WARN [Thread-21] org.apache.hadoop.mapred.LocalJobRunner - job_local923591174_0001
java.lang.Exception: java.io.FileNotFoundException: /opt/xdrive/druid-0.9.2/var/druid/hadoop-tmp/2017-06-30T080914.274Z_9baa0ee460f04b3f9d52624b0f6f7975/20140821T000000.000Z_20140822T000000.000Z/partitions.json (No space left on device)
at org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462) ~[hadoop-mapreduce-client-common-2.3.0.jar:?]
at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:529) [hadoop-mapreduce-client-common-2.3.0.jar:?]
Caused by: java.io.FileNotFoundException: /opt/xdrive/druid-0.9.2/var/druid/hadoop-tmp/my_datasource_name/2017-06-30T080914.274Z_9baa0ee460f04b3f9d52624b0f6f7975/20140821T000000.000Z_20140822T000000.000Z/partitions.json (No space left on device)
at java.io.FileOutputStream.open(Native Method) ~[?:1.7.0_121]
at java.io.FileOutputStream.<init>(FileOutputStream.java:221) ~[?:1.7.0_121]
at org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:206) ~[hadoop-common-2.3.0.jar:?]
at org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:202) ~[hadoop-common-2.3.0.jar:?]
at org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:265) ~[hadoop-common-2.3.0.jar:?]
at org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:252) ~[hadoop-common-2.3.0.jar:?]
at org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSOutputSummer.<init>(ChecksumFileSystem.java:384) ~[hadoop-common-2.3.0.jar:?]
at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:443) ~[hadoop-common-2.3.0.jar:?]
at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424) ~[hadoop-common-2.3.0.jar:?]
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:907) ~[hadoop-common-2.3.0.jar:?]
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:888) ~[hadoop-common-2.3.0.jar:?]
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:785) ~[hadoop-common-2.3.0.jar:?]
at io.druid.indexer.Utils.makePathAndOutputStream(Utils.java:70) ~[druid-indexing-hadoop-0.9.2.jar:0.9.2]
at io.druid.indexer.DetermineHashedPartitionsJob$DetermineCardinalityReducer.reduce(DetermineHashedPartitionsJob.java:328) ~[druid-indexing-hadoop-0.9.2.jar:0.9.2]
at io.druid.indexer.DetermineHashedPartitionsJob$DetermineCardinalityReducer.reduce(DetermineHashedPartitionsJob.java:299) ~[druid-indexing-hadoop-0.9.2.jar:0.9.2]
at org.apache.hadoop.mapreduce.Reducer.run(Reducer.java:171) ~[hadoop-mapreduce-client-core-2.3.0.jar:?]
at io.druid.indexer.DetermineHashedPartitionsJob$DetermineCardinalityReducer.run(DetermineHashedPartitionsJob.java:351) ~[druid-indexing-hadoop-0.9.2.jar:0.9.2]
at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:627) ~[hadoop-mapreduce-client-core-2.3.0.jar:?]
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389) ~[hadoop-mapreduce-client-core-2.3.0.jar:?]
at org.apache.hadoop.mapred.LocalJobRunner$Job$ReduceTaskRunnable.run(LocalJobRunner.java:319) ~[hadoop-mapreduce-client-common-2.3.0.jar:?]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) ~[?:1.7.0_121]
at java.util.concurrent.FutureTask.run(FutureTask.java:262) ~[?:1.7.0_121]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) ~[?:1.7.0_121]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) ~[?:1.7.0_121]
at java.lang.Thread.run(Thread.java:745) ~[?:1.7.0_121]
2017-06-30T08:10:02,193 INFO [communication thread] org.apache.hadoop.mapred.LocalJobRunner - reduce > reduce
2017-06-30T08:10:02,214 INFO [communication thread] org.apache.hadoop.mapred.LocalJobRunner - reduce > reduce