2020-04-08 22:37:32,063 - ERROR [streaming-job-executor-0:i.c.c.e.s.s.f.StreamingBatchSinkFunction@112] - Error writing to sink GCS Multi File for the batch for time 1586385430000.
java.lang.IllegalArgumentException: GCS Multi File has no outputs. Please check that the sink calls addOutput at some point.
at io.cdap.cdap.etl.spark.batch.SparkBatchSinkFactory.writeFromRDD(SparkBatchSinkFactory.java:85) ~[hydrator-spark-core2_2.11-6.1.1.jar:na]
at io.cdap.cdap.etl.spark.streaming.function.StreamingBatchSinkFunction.call(StreamingBatchSinkFunction.java:100) [hydrator-spark-core2_2.11-6.1.1.jar:na]
at io.cdap.cdap.etl.spark.streaming.function.StreamingBatchSinkFunction.call(StreamingBatchSinkFunction.java:51) [hydrator-spark-core2_2.11-6.1.1.jar:na]
at io.cdap.cdap.etl.spark.Compat$1.call(Compat.java:65) [hydrator-spark-core2_2.11-6.1.1.jar:na]
at io.cdap.cdap.etl.spark.Compat$1.call(Compat.java:62) [hydrator-spark-core2_2.11-6.1.1.jar:na]
at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$2.apply(JavaDStreamLike.scala:280) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$2.apply(JavaDStreamLike.scala:280) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:51) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:51) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:51) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:416) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:50) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:50) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:50) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at scala.util.Try$.apply(Try.scala:192) [scala-library-2.11.8.jar:na]
at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:257) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:257) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:257) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58) [scala-library-2.11.8.jar:na]
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:256) [spark-streaming_2.11-2.3.4.jar:2.3.4]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [na:1.8.0_242]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [na:1.8.0_242]
at java.lang.Thread.run(Thread.java:748) [na:1.8.0_242]
2020-04-08 22:37:32,080 - DEBUG [spark-listener-group-shared:i.c.c.a.r.s.AbstractSparkExecutionContext@154] - Spark program=program:default.test_avro_pubsub.-SNAPSHOT.spark.DataStreamsSparkStreaming, runId=9f5c6eb4-79e8-11ea-b920-a2fea9e6d990, jobId=3 starts without transaction
2020-04-08 22:37:32,080 - DEBUG [spark-listener-group-shared:i.c.c.a.r.s.SparkTransactionHandler@110] - Spark job started: JobTransaction{jobId=3, stageIds=[4], transaction=null}
2020-04-08 22:37:32,125 - DEBUG [spark-listener-group-shared:i.c.c.a.r.s.SparkTransactionHandler@144] - Spark job ended: JobTransaction{jobId=3, stageIds=[4], transaction=null}