--
You received this message because you are subscribed to the Google Groups "Confluent Platform" group.
To unsubscribe from this group and stop receiving emails from it, send an email to confluent-platform+unsub...@googlegroups.com.
To post to this group, send email to confluent-platform@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/confluent-platform/78324b49-1fdf-41f2-8dc8-7a3b49e205fb%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
Thanks for responding. s3.part.size has been left at the default104857600. Currently we are outputting to around 10 partitions, but it can be as high as 30. It just depends on the data. However it seems like 10 is more standard.
Jeff--
You received this message because you are subscribed to the Google Groups "Confluent Platform" group.
To unsubscribe from this group and stop receiving emails from it, send an email to confluent-platform+unsub...@googlegroups.com.
To post to this group, send email to confluent-platform@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/confluent-platform/e6ae9896-7b52-45b4-8761-207aea95a594%40googlegroups.com.
To post to this group, send email to confluent...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/confluent-platform/e6ae9896-7b52-45b4-8761-207aea95a594%40googlegroups.com.
[2017-06-07 16:29:37,051] INFO Opening record writer for: the_topic/the_topic/the_partition=XWS/the_topic+0+0000120920000.avro (io.confluent.connect.s3.format.avro.AvroRecordWriterProvider:66)
[2017-06-07 16:29:37,056] INFO WorkerSinkTask{id=s3-sink-lin-0} Committing offsets (org.apache.kafka.connect.runtime.WorkerSinkTask:272)
[2017-06-07 16:29:37,056] WARN Commit of WorkerSinkTask{id=s3-sink-lin-0} offsets timed out (org.apache.kafka.connect.runtime.WorkerSinkTask:172)
[2017-06-07 16:29:37,404] INFO Opening record writer for: the_topic/the_topic/the_partition=XWO/the_topic+0+0000120928433.avro (io.confluent.connect.s3.format.avro.AvroRecordWriterProvider:66)
[2017-06-07 16:29:37,423] INFO Opening record writer for: the_topic/the_topic/the_partition=BZA/the_topic+0+0000120928773.avro (io.confluent.connect.s3.format.avro.AvroRecordWriterProvider:66)
[2017-06-07 16:29:37,546] INFO Opening record writer for: the_topic/the_topic/the_partition=BZS/the_topic+0+0000120932021.avro (io.confluent.connect.s3.format.avro.AvroRecordWriterProvider:66)
[2017-06-07 16:29:37,897] INFO Opening record writer for: the_topic/the_topic/the_partition=XWW/the_topic+0+0000120941077.avro (io.confluent.connect.s3.format.avro.AvroRecordWriterProvider:66)
[2017-06-07 16:29:39,133] INFO Opening record writer for: the_topic/the_topic/the_partition=XWA/the_topic+0+0000120973464.avro (io.confluent.connect.s3.format.avro.AvroRecordWriterProvider:66)
[2017-06-07 16:29:39,801] INFO Opening record writer for: the_topic/the_topic/the_partition=BZO/the_topic+0+0000120990635.avro (io.confluent.connect.s3.format.avro.AvroRecordWriterProvider:66)
[2017-06-07 16:29:40,915] INFO Starting commit and rotation for topic partition the_topic-0 with start offset {the_partition=XWA=120973464, the_partition=XWO=120928433, the_partition=BZO=120990635, the_partition=XWS=120920000, the_partition=BZS=120932021, the_partition=BZA=120928773, the_partition=XWW=120941077} (io.confluent.connect.s3.TopicPartitionWriter:194)
[2017-06-07 16:30:38,958] INFO Files committed to S3. Target commit offset for the_topic-0 is 121020000 (io.confluent.connect.s3.TopicPartitionWriter:341)
How do we tell if the tasks are truly getting utilized and running in parallel?
Also, we did try an put out another worker, however we ran into two problems. BTW, this is two workers on the same node. Unfortunately we aren't in a position to have a clustered setup at this time. The first problem we encountered is that the same s3 connector was running on both workers and attempting to pull the same offsets at the same time therefore duplicating the data in s3. Generally one of the workers would fail and then both would advance to the next offset. We stopped and deleted the second worker at this point. We then tried to add a second worker again. This time the second worker just sat idle. We weren't sure when work is distributed to multiple workers, is it round robin? Or is the second worker only utilized when there is significant stain on the first worker?
Thanks again for all your help,
Jeff