I went ahead and downloaded Hadoop 1.2.1 (locally) and was able to compile the wordcount.cmr. But when i wanted to submit the job, I get the following error:
Analyzing job [count words]...
Executing jobs serially
Executing job [count words]....
Setting partitioner: com.linkedin.cubert.plan.physical.CubertPartitioner
15/06/22 11:31:10 INFO util.NativeCodeLoader: Loaded the native-hadoop library
15/06/22 11:31:10 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.
15/06/22 11:31:10 INFO input.FileInputFormat: Total input paths to process : 1
15/06/22 11:31:10 INFO util.MapRedUtil: Total input paths to process : 1
15/06/22 11:31:10 WARN snappy.LoadSnappy: Snappy native library not loaded
15/06/22 11:31:10 WARN mapred.LocalJobRunner: LocalJobRunner does not support symlinking into current working dir.
15/06/22 11:31:11 INFO mapred.LocalJobRunner: Waiting for map tasks
15/06/22 11:31:11 INFO mapred.LocalJobRunner: Starting task: attempt_local517226951_0001_m_000000_0
15/06/22 11:31:11 INFO util.ProcessTree: setsid exited with exit code 0
15/06/22 11:31:11 INFO mapred.Task: Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@75c2b15c
15/06/22 11:31:11 INFO mapred.MapTask: Processing split: file:[path to release]/words.txt:0+27294 (class org.apache.hadoop.mapreduce.lib.input.FileSplit) [0]
15/06/22 11:31:11 INFO mapred.MapTask: io.sort.mb = 100
15/06/22 11:31:11 INFO mapred.MapTask: data buffer = 79691776/99614720
15/06/22 11:31:11 INFO mapred.MapTask: record buffer = 262144/327680
Mapper init ----------------------------------
Executed operator chain for 1 block(s) in 254 ms
Mapper complete ----------------------------------
MemoryStats: #GC calls: 2 Total GC Time: 172 ms
15/06/22 11:31:11 INFO mapred.MapTask: Starting flush of map output
15/06/22 11:31:11 INFO mapred.MapTask: Finished spill 0
15/06/22 11:31:11 INFO mapred.Task: Task:attempt_local517226951_0001_m_000000_0 is done. And is in the process of commiting
15/06/22 11:31:11 INFO mapred.LocalJobRunner:
15/06/22 11:31:11 INFO mapred.Task: Task 'attempt_local517226951_0001_m_000000_0' done.
15/06/22 11:31:11 INFO mapred.LocalJobRunner: Finishing task: attempt_local517226951_0001_m_000000_0
15/06/22 11:31:11 INFO mapred.LocalJobRunner: Map task executor complete.
15/06/22 11:31:11 INFO mapred.Task: Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@7d5a3c12
15/06/22 11:31:11 INFO mapred.LocalJobRunner:
15/06/22 11:31:11 INFO mapred.Merger: Merging 1 sorted segments
15/06/22 11:31:12 INFO mapred.Merger: Down to the last merge-pass, with 1 segments left of total size: 20639 bytes
15/06/22 11:31:12 INFO mapred.LocalJobRunner:
Reducer init --------------------------------
15/06/22 11:31:12 WARN mapred.LocalJobRunner: job_local517226951_0001
java.lang.NoClassDefFoundError: com/google/common/base/Charsets
at org.apache.pig.impl.util.StorageUtil.putField(StorageUtil.java:185)
at org.apache.pig.impl.util.StorageUtil.putField(StorageUtil.java:116)
at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigTextOutputFormat$PigLineRecordWriter.write(PigTextOutputFormat.java:68)
at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigTextOutputFormat$PigLineRecordWriter.write(PigTextOutputFormat.java:44)
at org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.write(ReduceTask.java:586)
at org.apache.hadoop.mapreduce.TaskInputOutputContext.write(TaskInputOutputContext.java:80)
at com.linkedin.cubert.plan.physical.CubertReducer$ReduceContext.write(CubertReducer.java:144)
at com.linkedin.cubert.io.text.TextBlockWriter.write(TextBlockWriter.java:49)
at com.linkedin.cubert.plan.physical.CubertReducer.run(CubertReducer.java:84)
at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:649)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:418)
at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:398)
Caused by: java.lang.ClassNotFoundException: com.google.common.base.Charsets
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
... 12 more
Exception in thread "main" java.lang.InterruptedException: Job count words failed!
at com.linkedin.cubert.plan.physical.JobExecutor.run(JobExecutor.java:160)
at com.linkedin.cubert.plan.physical.ExecutorService.executeJob(ExecutorService.java:253)
at com.linkedin.cubert.plan.physical.ExecutorService.executeJobId(ExecutorService.java:219)
at com.linkedin.cubert.plan.physical.ExecutorService.execute(ExecutorService.java:163)
at com.linkedin.cubert.ScriptExecutor.execute(ScriptExecutor.java:385)
at com.linkedin.cubert.ScriptExecutor.main(ScriptExecutor.java:575)