{"type":"MAP_ATTEMPT_FAILED","event":{"org.apache.hadoop.mapreduce.jobhistory.TaskAttemptUnsuccessfulCompletion":{"taskid":"task_1456259383224_0001_m_000000","taskType":"MAP","attemptId":"attempt_1456259383224_0001_m_000000_0","finishTime":1456260106329,"hostname":"ip-10-206-40-223.dqa.capitalone.com","port":8041,"rackname":"/default-rack","status":"FAILED","error":"Error: cascading.tap.TapException: No suitable driver found for jdbc:redshift://cloudanal.cbljilhgawom.us-east-1.redshift.amazonaws.com:8080/dev (SQL error code: 0) opening connection: jdbc:redshift://cloudanal.cbljilhgawom.us-east-1.redshift.amazonaws.com:8080/dev\n\tat jdbc.JDBCTap.createConnection(JDBCTap.java:428)\n\tat jdbc.JDBCTap.resourceExists(JDBCTap.java:540)\n\tat jdbc.JDBCTap.createResource(JDBCTap.java:492)\n\tat jdbc.RedshiftTap.createResource(RedshiftTap.java:145)\n\tat jdbc.RedshiftTap.createResource(RedshiftTap.java:24)\n\tat cascading.tap.Tap.createResource(Tap.java:428)\n\tat jdbc.RedshiftTap.openForWrite(RedshiftTap.java:119)\n\tat jdbc.RedshiftTap.openForWrite(RedshiftTap.java:24)\n\tat cascading.flow.stream.SinkStage.prepare(SinkStage.java:60)\n\tat cascading.flow.stream.StreamGraph.prepare(StreamGraph.java:167)\n\tat cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:110)\n\tat org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:455)\n\tat org.apache.hadoop.mapred.MapTask.run(MapTask.java:344)\n\tat org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:172)\n\tat java.security.AccessController.doPrivileged(Native Method)\n\tat javax.security.auth.Subject.doAs(Subject.java:415)\n\tat org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)\n\tat org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:166)\nCaused by: java.sql.SQLException: No suitable driver found for jdbc:redshift://cloudanal.cbljilhgawom.us-east-1.redshift.amazonaws.com:8080/dev\n\tat java.sql.DriverManager.getConnection(DriverManager.java:596)\n\tat java.sql.DriverManager.getConnection(DriverManager.java:215)\n\tat jdbc.JDBCTap.createConnection(JDBCTap.java:410)\n\t... 17 more\n","counters":{"org.apache.hadoop.mapreduce.jobhistory.JhCounters":{"name":"COUNTERS","groups":[{"name":"org.apache.hadoop.mapreduce.FileSystemCounter","displayName":"File System Counters","counts":[{"name":"FILE_BYTES_READ","displayName":"FILE: Number of bytes read","value":0},{"name":"FILE_BYTES_WRITTEN","displayName":"FILE: Number of bytes written","value":130669},{"name":"FILE_READ_OPS","displayName":"FILE: Number of read operations","value":0},{"name":"FILE_LARGE_READ_OPS","displayName":"FILE: Number of large read operations","value":0},{"name":"FILE_WRITE_OPS","displayName":"FILE: Number of write operations","value":0},{"name":"HDFS_BYTES_READ","displayName":"HDFS: Number of bytes read","value":276},{"name":"HDFS_BYTES_WRITTEN","displayName":"HDFS: Number of bytes written","value":0},{"name":"HDFS_READ_OPS","displayName":"HDFS: Number of read operations","value":1},{"name":"HDFS_LARGE_READ_OPS","displayName":"HDFS: Number of large read operations","value":0},{"name":"HDFS_WRITE_OPS","displayName":"HDFS: Number of write operations","value":0},{"name":"S3_BYTES_READ","displayName":"S3: Number of bytes read","value":0},{"name":"S3_BYTES_WRITTEN","displayName":"S3: Number of bytes written","value":0},{"name":"S3_READ_OPS","displayName":"S3: Number of read operations","value":0},{"name":"S3_LARGE_READ_OPS","displayName":"S3: Number of large read operations","value":0},{"name":"S3_WRITE_OPS","displayName":"S3: Number of write operations","value":0}]},{"name":"org.apache.hadoop.mapreduce.TaskCounter","displayName":"Map-Reduce Framework","counts":[{"name":"MAP_INPUT_RECORDS","displayName":"Map input records","value":0},{"name":"MAP_OUTPUT_RECORDS","displayName":"Map output records","value":0},{"name":"SPLIT_RAW_BYTES","displayName":"Input split bytes","value":276},{"name":"SPILLED_RECORDS","displayName":"Spilled Records","value":0},{"name":"FAILED_SHUFFLE","displayName":"Failed Shuffles","value":0},{"name":"MERGED_MAP_OUTPUTS","displayName":"Merged Map outputs","value":0},{"name":"GC_TIME_MILLIS","displayName":"GC time elapsed (ms)","value":225},{"name":"CPU_MILLISECONDS","displayName":"CPU time spent (ms)","value":3920},{"name":"PHYSICAL_MEMORY_BYTES","displayName":"Physical memory (bytes) snapshot","value":281309184},{"name":"VIRTUAL_MEMORY_BYTES","displayName":"Virtual memory (bytes) snapshot","value":2055344128},{"name":"COMMITTED_HEAP_BYTES","displayName":"Total committed heap usage (bytes)","value":298844160}]},{"name":"cascading.flow.SliceCounters","displayName":"cascading.flow.SliceCounters","counts":[{"name":"Process_Begin_Time","displayName":"Process_Begin_Time","value":1456260106012}]},{"name":"org.apache.hadoop.mapreduce.lib.input.FileInputFormatCounter","displayName":"File Input Format Counters ","counts":[{"name":"BYTES_READ","displayName":"Bytes Read","value":0}]},{"name":"org.apache.hadoop.mapreduce.lib.output.FileOutputFormatCounter","displayName":"File Output Format Counters ","counts":[{"name":"BYTES_WRITTEN","displayName":"Bytes Written","value":0}]}]}},"clockSplits":[13393,2,1,2,1,2,2,1,2,1,2,2],"cpuUsages":[3920,0,0,0,0,0,0,0,0,0,0,0],"vMemKbytes":[2007171,2007171,2007172,2007171,2007172,2007171,2007172,2007171,2007172,2007172,2007171,2007172],"physMemKbytes":[274716,274716,274716,274715,274716,274715,274716,274715,274716,274716,274715,274716]}}}
From: jnkrish
Sent: February 23, 2016 8:09:48pm PST
To: cascading-user
Subject: Redshift issue
--
You received this message because you are subscribed to the Google Groups "cascading-user" group.
To unsubscribe from this group and stop receiving emails from it, send an email to cascading-use...@googlegroups.com.
To post to this group, send email to cascadi...@googlegroups.com.
Visit this group at https://groups.google.com/group/cascading-user.
To view this discussion on the web visit https://groups.google.com/d/msgid/cascading-user/f31fdd2b-2c72-4283-bb4a-d196d3dc1012%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
Error: cascading.flow.stream.DuctException: failed opening sink\n\tat cascading.flow.stream.SinkStage.prepare(SinkStage.java:70)\n\tat cascading.flow.stream.StreamGraph.prepare(StreamGraph.java:167)\n\tat cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:110)\n\tat org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:455)\n\tat org.apache.hadoop.mapred.MapTask.run(MapTask.java:344)\n\tat org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:172)\n\tat java.security.AccessController.doPrivileged(Native Method)\n\tat javax.security.auth.Subject.doAs(Subject.java:415)\n\tat org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)\n\tat org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:166)\nCaused by: java.io.IOException: SQL error code: 0 executing update statement: CREATE TABLE TEST123 ( id varchar(100), val varchar(100) ) \n\tat jdbc.JDBCUtil.executeUpdate(JDBCUtil.java:112)\n\tat jdbc.JDBCUtil.createTableIfNotExists(JDBCUtil.java:90)\n\tat jdbc.JDBCTap.createResource(JDBCTap.java:499)\n\tat jdbc.RedshiftTap.createResource(RedshiftTap.java:145)\n\tat jdbc.RedshiftTap.createResource(RedshiftTap.java:24)\n\tat cascading.tap.Tap.createResource(Tap.java:428)\n\tat jdbc.RedshiftTap.openForWrite(RedshiftTap.java:119)\n\tat jdbc.RedshiftTap.openForWrite(RedshiftTap.java:24)\n\tat cascading.flow.stream.SinkStage.prepare(SinkStage.java:60)\n\t... 9 more\nCaused by: org.postgresql.util.PSQLException: ERROR: Relation \"test123\" already exists\n\tat org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2198)\n\tat org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:1927)\n\tat org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:255)\n\tat org.postgresql.jdbc2.AbstractJdbc2Statement.execute(AbstractJdbc2Statement.java:561)\n\tat org.postgresql.jdbc2.AbstractJdbc2Statement.executeWithFlags(AbstractJdbc2Statement.java:405)\n\tat org.postgresql.jdbc2.AbstractJdbc2Statement.executeUpdate(AbstractJdbc2Statement.java:333)\n\tat jdbc.JDBCUtil.executeUpdate(JDBCUtil.java:105)\n\t... 17 more\n"
{"type":"TASK_FAILED","event":{"org.apache.hadoop.mapreduce.jobhistory.TaskFailed":{"taskid":"task_1456784336418_0003_m_000001","taskType":"MAP","finishTime":1456787411041,"error":", Error: cascading.flow.stream.DuctException: failed opening sink\n\tat cascading.flow.stream.SinkStage.prepare(SinkStage.java:70)\n\tat cascading.flow.stream.StreamGraph.prepare(StreamGraph.java:167)\n\tat cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:110)\n\tat org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:455)\n\tat org.apache.hadoop.mapred.MapTask.run(MapTask.java:344)\n\tat org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:172)\n\tat java.security.AccessController.doPrivileged(Native Method)\n\tat javax.security.auth.Subject.doAs(Subject.java:415)\n\tat org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)\n\tat org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:166)\nCaused by: java.io.IOException: SQL error code: 0 executing update statement: CREATE TABLE TEST123 ( id varchar(30), val varchar(30) ) DISTKEY (id) SORTKEY (id) \n\tat jdbc.JDBCUtil.executeUpdate(JDBCUtil.java:115)\n\tat jdbc.JDBCUtil.createTableIfNotExists(JDBCUtil.java:93)\n\tat jdbc.JDBCTap.createResource(JDBCTap.java:499)\n\tat jdbc.RedshiftTap.createResource(RedshiftTap.java:145)\n\tat jdbc.RedshiftTap.createResource(RedshiftTap.java:24)\n\tat cascading.tap.Tap.createResource(Tap.java:428)\n\tat jdbc.RedshiftTap.openForWrite(RedshiftTap.java:119)\n\tat jdbc.RedshiftTap.openForWrite(RedshiftTap.java:24)\n\tat cascading.flow.stream.SinkStage.prepare(SinkStage.java:60)\n\t... 9 more\nCaused by: org.postgresql.util.PSQLException: ERROR: Relation \"test123\" already exists\n\tat org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2198)\n\tat org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:1927)\n\tat org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:255)\n\tat org.postgresql.jdbc2.AbstractJdbc2Statement.execute(AbstractJdbc2Statement.java:561)\n\tat org.postgresql.jdbc2.AbstractJdbc2Statement.executeWithFlags(AbstractJdbc2Statement.java:405)\n\tat org.postgresql.jdbc2.AbstractJdbc2Statement.executeUpdate(AbstractJdbc2Statement.java:333)\n\tat jdbc.JDBCUtil.executeUpdate(JDBCUtil.java:108)\n\t... 17 more\n","failedDueToAttempt":{"string":"attempt_1456784336418_0003_m_000001_3"},"status":"FAILED","counters":{"org.apache.hadoop.mapreduce.jobhistory.JhCounters":{"name":"COUNTERS","groups":[{"name":"org.apache.hadoop.mapreduce.TaskCounter","displayName":"Map-Reduce Framework","counts":[{"name":"CPU_MILLISECONDS","displayName":"CPU time spent (ms)","value":0},{"name":"PHYSICAL_MEMORY_BYTES","displayName":"Physical memory (bytes) snapshot","value":0},{"name":"VIRTUAL_MEMORY_BYTES","displayName":"Virtual memory (bytes) snapshot","value":0}]}]}}}}}
...