[2018-03-22 16:38:53,833] INFO KafkaDatabaseHistory Consumer config: {enable.auto.commit=false, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer,
group.id=6mysql-source-connector-dbhistory, auto.offset.reset=earliest,
session.timeout.ms=10000, bootstrap.servers=kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9092,kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9093,kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9094,
client.id=6mysql-source-connector-dbhistory, key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, fetch.min.bytes=1} (io.debezium.relational.history.KafkaDatabaseHistory:163)
[2018-03-22 16:38:53,833] INFO KafkaDatabaseHistory Consumer config: {enable.auto.commit=false, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer,
group.id=6mysql-source-connector-dbhistory, auto.offset.reset=earliest,
session.timeout.ms=10000, bootstrap.servers=kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9092,kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9093,kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9094,
client.id=6mysql-source-connector-dbhistory, key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, fetch.min.bytes=1} (io.debezium.relational.history.KafkaDatabaseHistory:163)
[2018-03-22 16:38:53,833] INFO KafkaDatabaseHistory Producer config: {bootstrap.servers=kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9092,kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9093,kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9094, value.serializer=org.apache.kafka.common.serialization.StringSerializer, buffer.memory=1048576, retries=1, key.serializer=org.apache.kafka.common.serialization.StringSerializer,
client.id=6mysql-source-connector-dbhistory,
linger.ms=0, batch.size=32768,
max.block.ms=10000, acks=1} (io.debezium.relational.history.KafkaDatabaseHistory:164)
[2018-03-22 16:38:53,833] INFO KafkaDatabaseHistory Producer config: {bootstrap.servers=kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9092,kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9093,kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9094, value.serializer=org.apache.kafka.common.serialization.StringSerializer, buffer.memory=1048576, retries=1, key.serializer=org.apache.kafka.common.serialization.StringSerializer,
client.id=6mysql-source-connector-dbhistory,
linger.ms=0, batch.size=32768,
max.block.ms=10000, acks=1} (io.debezium.relational.history.KafkaDatabaseHistory:164)
[2018-03-22 16:38:53,833] INFO ProducerConfig values:
acks = 1
batch.size = 32768
bootstrap.servers = [kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9092, kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9093, kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9094]
buffer.memory = 1048576
client.id = 6mysql-source-connector-dbhistory
compression.type = none
enable.idempotence = false
interceptor.classes = null
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
retries = 1
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
ssl.endpoint.identification.algorithm = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLS
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:238)
[2018-03-22 16:38:53,835] INFO Kafka version : 1.0.1 (org.apache.kafka.common.utils.AppInfoParser:109)
[2018-03-22 16:38:53,835] INFO Kafka commitId : unknown (org.apache.kafka.common.utils.AppInfoParser:110)
[2018-03-22 16:38:54,477] INFO AdminClientConfig values:
bootstrap.servers = [kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9092, kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9093, kafka-d7cf38e5682303e4.us-west-2.aws.confluent.cloud:9094]
client.id = 6mysql-source-connector-dbhistory
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
receive.buffer.bytes = 65536
retries = 1
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
ssl.endpoint.identification.algorithm = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLS
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:238)
[2018-03-22 16:38:54,478] WARN The configuration 'value.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:246)
[2018-03-22 16:38:54,478] WARN The configuration 'batch.size' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:246)
[2018-03-22 16:38:54,478] WARN The configuration '
max.block.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:246)
[2018-03-22 16:38:54,478] WARN The configuration 'acks' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:246)
[2018-03-22 16:38:54,479] WARN The configuration 'buffer.memory' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:246)
[2018-03-22 16:38:54,479] WARN The configuration 'key.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:246)
[2018-03-22 16:38:54,479] WARN The configuration '
linger.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:246)
[2018-03-22 16:38:54,479] INFO Kafka version : 1.0.1 (org.apache.kafka.common.utils.AppInfoParser:109)
[2018-03-22 16:38:54,479] INFO Kafka commitId : unknown (org.apache.kafka.common.utils.AppInfoParser:110)
[2018-03-22 16:38:57,484] INFO Stopping MySQL connector task (io.debezium.connector.mysql.MySqlConnectorTask:239)
[2018-03-22 16:38:57,484] INFO Stopping MySQL connector task (io.debezium.connector.mysql.MySqlConnectorTask:239)
[2018-03-22 16:38:57,485] INFO WorkerSourceTask{id=6mysql-source-connector-0} Committing offsets (org.apache.kafka.connect.runtime.WorkerSourceTask:306)
[2018-03-22 16:38:57,485] INFO WorkerSourceTask{id=6mysql-source-connector-0} flushing 0 outstanding messages for offset commit (org.apache.kafka.connect.runtime.WorkerSourceTask:323)
[2018-03-22 16:38:57,485] ERROR WorkerSourceTask{id=6mysql-source-connector-0} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask:172)