I modified the index_hadoop JSON to correctly include hadoopCoordinates. Now i dont see above exception. First M/R job completes, Immediately after that i see below exception
2014-04-22 16:56:50,560 WARN [task-runner-0] org.apache.hadoop.mapred.JobClient - No job jar file set. User classes may not be found. See JobConf(Class) or JobConf#setJar(String).
2014-04-22 16:56:50,567 INFO [task-runner-0] org.apache.hadoop.mapreduce.lib.input.FileInputFormat - Total input paths to process : 1
2014-04-22 16:56:50,585 WARN [task-runner-0] org.apache.hadoop.util.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2014-04-22 16:56:50,586 WARN [task-runner-0] org.apache.hadoop.io.compress.snappy.LoadSnappy - Snappy native library not loaded
2014-04-22 16:56:51,246 INFO [task-runner-0] io.druid.indexer.IndexGeneratorJob - Job wikipedia-index-generator-Optional.of([2013-08-31T00:00:00.000Z/2013-09-01T00:00:00.000Z]) submitted, status available at http://hadoop-server-249608.slc01.dev.ebayc3.com:50030/jobdetails.jsp?jobid=job_201404221645_0004
2014-04-22 16:56:51,246 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Running job: job_201404221645_0004
2014-04-22 16:56:52,250 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - map 0% reduce 0%
2014-04-22 16:57:01,273 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - map 100% reduce 0%
2014-04-22 16:57:10,301 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - map 100% reduce 33%
2014-04-22 16:57:14,313 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - map 100% reduce 100%
2014-04-22 16:57:16,321 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Job complete: job_201404221645_0004
2014-04-22 16:57:16,343 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Counters: 29
2014-04-22 16:57:16,343 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Job Counters
2014-04-22 16:57:16,343 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Launched reduce tasks=1
2014-04-22 16:57:16,343 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - SLOTS_MILLIS_MAPS=10723
2014-04-22 16:57:16,343 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Total time spent by all reduces waiting after reserving slots (ms)=0
2014-04-22 16:57:16,344 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Total time spent by all maps waiting after reserving slots (ms)=0
2014-04-22 16:57:16,344 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Launched map tasks=1
2014-04-22 16:57:16,344 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Data-local map tasks=1
2014-04-22 16:57:16,344 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - SLOTS_MILLIS_REDUCES=12322
2014-04-22 16:57:16,345 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - File Output Format Counters
2014-04-22 16:57:16,345 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Bytes Written=0
2014-04-22 16:57:16,345 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - FileSystemCounters
2014-04-22 16:57:16,345 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - FILE_BYTES_READ=1872
2014-04-22 16:57:16,345 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - HDFS_BYTES_READ=1817
2014-04-22 16:57:16,346 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - FILE_BYTES_WRITTEN=256253
2014-04-22 16:57:16,346 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - HDFS_BYTES_WRITTEN=2867
2014-04-22 16:57:16,346 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - File Input Format Counters
2014-04-22 16:57:16,346 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Bytes Read=1675
2014-04-22 16:57:16,346 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Map-Reduce Framework
2014-04-22 16:57:16,347 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Map output materialized bytes=1872
2014-04-22 16:57:16,347 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Map input records=5
2014-04-22 16:57:16,347 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Reduce shuffle bytes=1872
2014-04-22 16:57:16,347 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Spilled Records=10
2014-04-22 16:57:16,347 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Map output bytes=1846
2014-04-22 16:57:16,348 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Total committed heap usage (bytes)=327155712
2014-04-22 16:57:16,348 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - CPU time spent (ms)=6850
2014-04-22 16:57:16,348 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Combine input records=0
2014-04-22 16:57:16,348 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - SPLIT_RAW_BYTES=142
2014-04-22 16:57:16,349 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Reduce input records=5
2014-04-22 16:57:16,349 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Reduce input groups=1
2014-04-22 16:57:16,349 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Combine output records=0
2014-04-22 16:57:16,349 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Physical memory (bytes) snapshot=425791488
2014-04-22 16:57:16,349 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Reduce output records=0
2014-04-22 16:57:16,349 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Virtual memory (bytes) snapshot=2027036672
2014-04-22 16:57:16,349 INFO [task-runner-0] org.apache.hadoop.mapred.JobClient - Map output records=5
2014-04-22 16:57:16,364 ERROR [task-runner-0] io.druid.indexing.overlord.ThreadPoolTaskRunner - Exception while running task[HadoopIndexTask{id=index_hadoop_wikipedia_2014-04-22T16:56:38.034Z, type=index_hadoop, dataSource=wikipedia}]
java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at io.druid.indexing.common.task.HadoopIndexTask.run(HadoopIndexTask.java:220)
at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:224)
at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:203)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IncompatibleClassChangeError: Found class org.apache.hadoop.mapreduce.Counter, but interface was expected
at io.druid.indexer.IndexGeneratorJob.run(IndexGeneratorJob.java:177)
at io.druid.indexer.JobHelper.runJobs(JobHelper.java:134)
at io.druid.indexer.HadoopDruidIndexerJob.run(HadoopDruidIndexerJob.java:80)
at io.druid.indexing.common.task.HadoopIndexTask$HadoopIndexGeneratorInnerProcessing.runTask(HadoopIndexTask.java:259)
... 11 more
2014-04-22 16:57:16,370 INFO [task-runner-0] io.druid.indexing.worker.executor.ExecutorLifecycle - Task completed with status: {
"id" : "index_hadoop_wikipedia_2014-04-22T16:56:38.034Z",
"status" : "FAILED",
"duration" : 31643
}
Any suggestions ?