Hi Raghu,
Got it to work. Yeah!
I decided to have a look at jobtracker logs after the following hive query failed:
hive> describe test1;
OK
name_ string from deserializer
age_ int from deserializer
memoizedserializedsize int from deserializer
dt string
Time taken: 0.117 seconds
Why does every element end with a _ ?hive> select name_ from test1;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks is set to 0 since there's no reduce operator
Starting Job = job_201208160854_27843, Tracking URL =
http://dacd002.us.msudev.noklab.net:50030/jobdetails.jsp?jobid=job_201208160854_27843Kill Command = /usr/lib/hadoop/bin/hadoop job -Dmapred.job.tracker=
dacd002.us.msudev.noklab.net:8021 -kill job_201208160854_27843
2012-08-30 20:15:55,103 Stage-1 map = 0%, reduce = 0%
2012-08-30 20:16:23,253 Stage-1 map = 100%, reduce = 100%
Ended Job = job_201208160854_27843 with errors
FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.MapRedTask
Looking at logCaused by: java.lang.NoClassDefFoundError: com/google/common/base/Function
at com.twitter.elephantbird.mapreduce.input.MultiInputFormat.determineFileFormat(MultiInputFormat.java:185)
at com.twitter.elephantbird.mapreduce.input.MultiInputFormat.createRecordReader(MultiInputFormat.java:87)
at com.twitter.elephantbird.mapreduce.input.RawMultiInputFormat.createRecordReader(RawMultiInputFormat.java:36)
at com.twitter.elephantbird.mapred.input.DeprecatedInputFormatWrapper$RecordReaderWrapper.<init>(DeprecatedInputFormatWrapper.java:230)
at com.twitter.elephantbird.mapred.input.DeprecatedInputFormatWrapper.getRecordReader(DeprecatedInputFormatWrapper.java:92)
at org.apache.hadoop.hive.ql.io.CombineHiveRecordReader.<init>(CombineHiveRecordReader.java:68)
Found this is in the guava-r06.jar, and I also thought having protobuf-java-2.3.0.jar would be a good idea too.
export HIVE_AUX_JARS_PATH="/nfs_home/jcroke/elephant/kevinweil-elephant-bird-4b28225/lib/elephant-bird-core-3.0.3-SNAPSHOT.jar:/nfs_home/jcroke/elephant/kevinweil-elephant-bird-4b28225/lib/elephant-bird-hive-3.0.3-SNAPSHOT.jar:/nfs_home/jcroke/elephant/kevinweil-elephant-bird-4b28225/lib/protobuf-java-2.3.0.jar:/nfs_home/jcroke/elephant/kevinweil-elephant-bird-4b28225/lib/guava-r06.jar"
[jcroke@daca2 kevinweil-elephant-bird-4b28225]$ hive
Hive history file=/tmp/jcroke/hive_job_log_jcroke_201208302103_1688217610.txt
hive> use posit;
OK
Time taken: 1.507 seconds
hive> select name_ from test1 where dt = '2012/07/16/12';
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks is set to 0 since there's no reduce operator
Starting Job = job_201208160854_28016, Tracking URL =
http://dacd002.us.msudev.noklab.net:50030/jobdetails.jsp?jobid=job_201208160854_28016Kill Command = /usr/lib/hadoop/bin/hadoop job -Dmapred.job.tracker=
dacd002.us.msudev.noklab.net:8021 -kill job_201208160854_28016
2012-08-30 21:04:19,435 Stage-1 map = 0%, reduce = 0%
2012-08-30 21:04:23,466 Stage-1 map = 100%, reduce = 0%
2012-08-30 21:04:24,475 Stage-1 map = 100%, reduce = 100%
Ended Job = job_201208160854_28016
OK
Jon
Jenn
David
Duncan
Time taken: 11.016 seconds
hive> select count(*) from test1;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapred.reduce.tasks=<number>
Starting Job = job_201208160854_28017, Tracking URL =
http://dacd002.us.msudev.noklab.net:50030/jobdetails.jsp?jobid=job_201208160854_28017Kill Command = /usr/lib/hadoop/bin/hadoop job -Dmapred.job.tracker=
dacd002.us.msudev.noklab.net:8021 -kill job_201208160854_28017
2012-08-30 21:04:39,394 Stage-1 map = 0%, reduce = 0%
2012-08-30 21:04:43,419 Stage-1 map = 100%, reduce = 0%
2012-08-30 21:04:52,473 Stage-1 map = 100%, reduce = 100%
Ended Job = job_201208160854_28017
OK
4
Time taken: 19.078 seconds
hive>
So I was missing some jar files and that was all.
Thanks for your Help : )
Jon