Show tables and Describe Table works fine. Those tables i had created on hive.
presto:default> show tables;
Table
----------------
pokes
productfind
productview
scenario
search
(5 rows)
Query 20131110_134848_00005_vwjwd, FINISHED, 1 node
Splits: 2 total, 1 done (50.00%)
0:00 [0 rows, 0B] [0 rows/s, 0B/s]
presto:default> describe pokes;
Column | Type | Null | Partition Key
--------+---------+------+---------------
foo | bigint | true | false
bar | varchar | true | false
(2 rows)
Query 20131110_134928_00007_vwjwd, FINISHED, 1 node
Splits: 2 total, 2 done (100.00%)
0:00 [2 rows, 131B] [5 rows/s, 368B/s]
But when i use select query and create table query, then exception is :
presto:default> select foo from pokes limit 5;
Query 20131110_135438_00008_vwjwd, FAILED, 1 node
Splits: 1 total, 0 done (0.00%)
0:00 [0 rows, 0B] [0 rows/s, 0B/s]
Query 20131110_135438_00008_vwjwd failed: java.io.IOException: Failed on local exception: java.io.IOException: Response is null.; Host Details : local host is: "tarun-ThinkPad-Edge-E430/127.0.1.1"; destination host is: "localhost":54310;
presto:default> CREATE TABLE pokesData (fooData INT, barData STRING);
Query 20131110_135528_00009_vwjwd failed: line 1:24: mismatched input '(' expecting AS
CREATE TABLE pokesData (fooData INT, barData STRING)
Kindly please let me know what i am doing wrong here.
Had this problem but this one was solved by providing this informations.
Hi,
Can you please tell me the syntax of how to add entry hadoop config files "hdfs-site.xml,mapred-site.xml,core-site.xml" in jvm.config of Presto
Can you please tell me the syntax of how to add entry hadoop config files "hdfs-site.xml,mapred-site.xml,core-site.xml" in jvm.config of Presto
I am not using cloudera distribution. I am using hadoop-1.2.1 and hive-0.11.
I am not using Cloudra's Hive release and i am using hive-0.10.0 and would like to go with existing setup.
Thanks for quick response . Still getting problem. I just execute create table query as :
presto:default> create table pokes (p STRING,q STRING);
Query 20131118_201139_00007_rwgzi failed: line 1:20: mismatched input '(' expecting AS
create table pokes (p STRING,q STRING)
Hi David,
query: SHOW TABLES -- working fine
query:DESCRIBE books -- working fine
following query not working - can you please guide me
Hadoop - 1.2.1
Hive - 0.11.0
presto-server: 0.52
presto:default> select * from books;
Query 20131121_025845_00004_qqe25, FAILED, 1 node
Splits: 1 total, 0 done (0.00%)
0:00 [0 rows, 0B] [0 rows/s, 0B/s]
Query 20131121_025845_00004_qqe25 failed: java.io.IOException: Failed on local exception: java.io.IOException: Broken pipe; Host Details : local host is: "ubuntu/192.168.56.101"; destination host is: "localhost":54310;
presto:default>
exception:
45_00004_qqe25.1
java.lang.RuntimeException: java.io.IOException: Failed on local exception: java.io.IOException: Broken pipe; Host Details : local host is: "ubuntu/192.168.56.101"; destination host is: "localhost":54310;
at com.google.common.base.Throwables.propagate(Throwables.java:160) ~[guava-15.0.jar:na]
at com.facebook.presto.hive.HiveSplitIterable$HiveSplitQueue.computeNext(HiveSplitIterable.java:433) ~[na:na]
at com.facebook.presto.hive.HiveSplitIterable$HiveSplitQueue.computeNext(HiveSplitIterable.java:392) ~[na:na]
at com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) ~[guava-15.0.jar:na]
at com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) ~[guava-15.0.jar:na]
at com.facebook.presto.execution.SqlStageExecution.startTasks(SqlStageExecution.java:463) [presto-main-0.52.jar:0.52]
at com.facebook.presto.execution.SqlStageExecution.access$300(SqlStageExecution.java:80) [presto-main-0.52.jar:0.52]
at com.facebook.presto.execution.SqlStageExecution$5.run(SqlStageExecution.java:435) [presto-main-0.52.jar:0.52]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) [na:1.7.0_45]
at java.util.concurrent.FutureTask.run(FutureTask.java:262) [na:1.7.0_45]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) [na:1.7.0_45]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) [na:1.7.0_45]
at java.lang.Thread.run(Thread.java:744) [na:1.7.0_45]
Caused by: java.io.IOException: Failed on local exception: java.io.IOException: Broken pipe; Host Details : local host is: "ubuntu/192.168.56.101"; destination host is: "localhost":54310;
at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:763) ~[na:na]
at org.apache.hadoop.ipc.Client.call(Client.java:1229) ~[na:na]
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202) ~[na:na]
at com.sun.proxy.$Proxy155.getListing(Unknown Source) ~[na:na]
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.7.0_45]
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) ~[na:1.7.0_45]
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.7.0_45]
at java.lang.reflect.Method.invoke(Method.java:606) ~[na:1.7.0_45]
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164) ~[na:na]
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83) ~[na:na]
at com.sun.proxy.$Proxy155.getListing(Unknown Source) ~[na:na]
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:441) ~[na:na]
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1526) ~[na:na]
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1509) ~[na:na]
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:406) ~[na:na]
at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1462) ~[na:na]
at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1502) ~[na:na]
at com.facebook.presto.hive.ForwardingFileSystem.listStatus(ForwardingFileSystem.java:298) ~[na:na]
at com.facebook.presto.hive.ForwardingFileSystem.listStatus(ForwardingFileSystem.java:298) ~[na:na]
at com.facebook.presto.hive.FileSystemWrapper$3.listStatus(FileSystemWrapper.java:146) ~[na:na]
at org.apache.hadoop.fs.FileSystem$4.<init>(FileSystem.java:1778) ~[na:na]
at org.apache.hadoop.fs.FileSystem.listLocatedStatus(FileSystem.java:1777) ~[na:na]
at org.apache.hadoop.fs.FileSystem.listLocatedStatus(FileSystem.java:1760) ~[na:na]
at com.facebook.presto.hive.util.AsyncRecursiveWalker$1.run(AsyncRecursiveWalker.java:58) ~[na:na]
at com.facebook.presto.hive.util.SuspendingExecutor$1.run(SuspendingExecutor.java:67) ~[na:na]
at com.facebook.presto.hive.util.BoundedExecutor.executeOrMerge(BoundedExecutor.java:82) ~[na:na]
at com.facebook.presto.hive.util.BoundedExecutor.access$000(BoundedExecutor.java:41) ~[na:na]
at com.facebook.presto.hive.util.BoundedExecutor$1.run(BoundedExecutor.java:53) ~[na:na]
... 3 common frames omitted
Caused by: java.io.IOException: Broken pipe
at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[na:1.7.0_45]
at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47) ~[na:1.7.0_45]
at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93) ~[na:1.7.0_45]
at sun.nio.ch.IOUtil.write(IOUtil.java:65) ~[na:1.7.0_45]
at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:487) ~[na:1.7.0_45]
at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[na:na]
at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:143) ~[na:na]
at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:153) ~[na:na]
at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:114) ~[na:na]
at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) ~[na:1.7.0_45]
at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) ~[na:1.7.0_45]
at java.io.DataOutputStream.flush(DataOutputStream.java:123) ~[na:1.7.0_45]
at org.apache.hadoop.ipc.Client$Connection$3.run(Client.java:897) ~[na:na]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) [na:1.7.0_45]
at java.util.concurrent.FutureTask.run(FutureTask.java:262) [na:1.7.0_45]
... 3 common frames omitted
2013-11-20T21:58:45.915-0500 DEBUG task-notification-1 com.facebook.presto.execution.TaskStateMachine Task 20131121_025845_00004_qqe25.0.0 is CANCELED
while trying to run SELECT command. Using hadoop-1.2.1 and hive-0.9.0.
Thanks in advance.
The variable connector.name (in presto-server-0.56/etc/catalog/hive.properties) was set to hive-cdh4. I changed it to hive-hadoop1 and now its working.
Being a beginner myself i don't really know what might be causing your issue, but please check once if you have set the connector name in presto-server-0.56/etc/catalog/hive.properties correctly.
Regards