Async worker died! ... clojure.lang.PersistentVector cannot be cast to class java.lang.String

146 views
Skip to first unread message

dgdesi...@gmail.com

unread,
Nov 20, 2019, 12:05:10 PM11/20/19
to DigitalPebble

Hi All!

I'm new to the crawler world and Java so please set me straight.

I've been successful injecting my target URLs but when I run the crawler local my async worker dies...

Googling has not helped me in this case.


Below is the print out of the situation.

Please let me know what other information I should include.


Any idea what's causing this?



...



storm jar target/dev1-0.1.jar org.apache.storm.flux.Flux --local --sleep 86400000 es-crawler.flux

Running: java -client -Ddaemon.name= -Dstorm.options= -Dstorm.home=/opt/apache-storm-1.2.3 -Dstorm.log.dir=/opt/apache-storm-1.2.3/logs -Djava.library.path=/usr/local/lib:/opt/local/lib:/usr/lib -Dstorm.conf.file= -cp /opt/apache-storm-1.2.3/*:/opt/apache-storm-1.2.3/lib/*:/opt/apache-storm-1.2.3/extlib/*:target/dev1-0.1.jar:/opt/apache-storm-1.2.3/conf:/opt/apache-storm-1.2.3/bin -Dstorm.jar=target/dev1-0.1.jar -Dstorm.dependency.jars= -Dstorm.dependency.artifacts={} org.apache.storm.flux.Flux --local --sleep 86400000 es-crawler.flux

███████╗██╗     ██╗   ██╗██╗  ██╗

██╔════╝██║     ██║   ██║╚██╗██╔╝

█████╗  ██║     ██║   ██║ ╚███╔╝

██╔══╝  ██║     ██║   ██║ ██╔██╗

██║     ███████╗╚██████╔╝██╔╝ ██╗

╚═╝     ╚══════╝ ╚═════╝ ╚═╝  ╚═╝

+-         Apache Storm        -+

+-  data FLow User eXperience  -+

Version: 1.2.3

Parsing file: /home/ubuntu/crawl/dev1/es-crawler.flux

809  [main] INFO  o.a.s.f.p.FluxParser - loading YAML from input stream...

814  [main] INFO  o.a.s.f.p.FluxParser - Not performing property substitution.

814  [main] INFO  o.a.s.f.p.FluxParser - Not performing environment variable substitution.

886  [main] INFO  o.a.s.f.p.FluxParser - Loading includes from resource: /crawler-default.yaml

891  [main] INFO  o.a.s.f.p.FluxParser - loading YAML from input stream...

895  [main] INFO  o.a.s.f.p.FluxParser - Not performing property substitution.

895  [main] INFO  o.a.s.f.p.FluxParser - Not performing environment variable substitution.


  fetcher.queue.mode: byHost

  fetcher.threads.per.queue: 1

  fetcher.threads.number: 10

  fetcher.max.urls.in.queues: -1

  fetcher.max.queue.size: -1

  fetcher.timeout.queue: -1

  fetcher.max.crawl.delay: 30

  fetcher.max.crawl.delay.force: false

  fetcher.server.delay.force: false

  fetcher.metrics.time.bucket.secs: 10

  fetcher.max.throttle.sleep: -1

  partition.url.mode: byHost

  urlbuffer.class: com.digitalpebble.stormcrawler.persistence.SimpleURLBuffer

  metadata.persist: [_redirTo, error.cause, error.source, isSitemap, isFeed]

  metadata.track.path: true

  metadata.track.depth: true

  http.agent.name: Anonymous Coward

  http.agent.version: '1.0'

  http.agent.description: built with StormCrawler 1.15

  http.agent.url: http://someorganization.com/

  http.agent.email: som...@someorganization.com

  http.accept.language: en-us,en-gb,en;q=0.7,*;q=0.3

  http.accept: text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8

  http.content.limit: -1

  http.store.headers: false

  http.timeout: 10000

  http.content.partial.as.trimmed: false

  http.robots.403.allow: true

  robots.noFollow.strict: true

  robots.cache.spec: maximumSize=10000,expireAfterWrite=6h

  robots.error.cache.spec: maximumSize=10000,expireAfterWrite=1h

  protocols: http,https,file

  http.protocol.implementation: com.digitalpebble.stormcrawler.protocol.httpclient.HttpProtocol

  https.protocol.implementation: com.digitalpebble.stormcrawler.protocol.httpclient.HttpProtocol

  file.protocol.implementation: com.digitalpebble.stormcrawler.protocol.file.FileProtocol

  selenium.implicitlyWait: 0

  selenium.pageLoadTimeout: -1

  selenium.setScriptTimeout: 0

  selenium.instances.num: 1

  selenium.capabilities: {takesScreenshot: false, loadImages: false, javascriptEnabled: true}

  selenium.delegated.protocol: com.digitalpebble.stormcrawler.protocol.httpclient.HttpProtocol

  jsoup.treat.non.html.as.error: true

  parser.emitOutlinks: true

  parser.emitOutlinks.max.per.page: -1

  track.anchors: true

  detect.mimetype: true

  detect.charset.maxlength: 10000

  sitemap.filter.hours.since.modified: -1

  sitemap.schedule.delay: -1

  sitemap.discovery: false

  scheduler.class: com.digitalpebble.stormcrawler.persistence.DefaultScheduler

  fetchInterval.default: 1440

  fetchInterval.fetch.error: 120

  fetchInterval.error: -1

  max.fetch.errors: 3

  status.updater.use.cache: true

  status.updater.cache.spec: maximumSize=10000,expireAfterAccess=1h

  status.updater.unit.round.date: SECOND

  indexer.url.fieldname: url

  indexer.text.fieldname: content

  indexer.text.maxlength: -1

  indexer.canonical.name: canonical

  indexer.md.mapping: [parse.title=title, parse.keywords=keywords, parse.description=description]

includes: null

name: null

spouts: []

streams: []

topologySource: null


929  [main] INFO  o.a.s.f.p.FluxParser - Loading includes from file: crawler-conf.yaml

929  [main] INFO  o.a.s.f.p.FluxParser - loading YAML from input stream...

933  [main] INFO  o.a.s.f.p.FluxParser - Not performing property substitution.

933  [main] INFO  o.a.s.f.p.FluxParser - Not performing environment variable substitution.

Configuration (interpreted): 

!!org.apache.storm.flux.model.TopologyDef

bolts: []

components: []

config:

  topology.workers: 1

  topology.message.timeout.secs: 300

  topology.max.spout.pending: 100

  topology.debug: false

  fetcher.threads.number: 50

  topology.worker.childopts: -Xmx2g -Djava.net.preferIPv4Stack=true

  topology.kryo.register: [com.digitalpebble.stormcrawler.Metadata]

  metadata.persist: [_redirTo, error.cause, error.source, isSitemap, isFeed]

  http.agent.name: Google-Read-Aloud

  http.agent.version: '0.1'

  http.agent.description: ''

  http.agent.url: https://support.google.com/webmasters/answer/1061943

  http.agent.email: ''

  http.content.limit: 65536

  parsefilters.config.file: parsefilters.json

  urlfilters.config.file: urlfilters.json

  fetchInterval.default: 1440

  fetchInterval.fetch.error: 120

  fetchInterval.error: -1

  textextractor.include.pattern: ['DIV[id="maincontent"]', 'DIV[itemprop="articleBody"]',

    ARTICLE]

  textextractor.exclude.tags: [STYLE, SCRIPT]

  indexer.url.fieldname: url

  indexer.text.fieldname: content

  indexer.canonical.name: canonical

  indexer.md.mapping: [parse.title=title, parse.keywords=keywords, parse.description=description,

    domain=domain]

  topology.metrics.consumer.register:

  - {class: org.apache.storm.metric.LoggingMetricsConsumer, parallelism.hint: 1}

includes: null

name: null

spouts: []

streams: []

topologySource: null


940  [main] INFO  o.a.s.f.p.FluxParser - Loading includes from file: es-conf.yaml

941  [main] INFO  o.a.s.f.p.FluxParser - loading YAML from input stream...

944  [main] INFO  o.a.s.f.p.FluxParser - Not performing property substitution.

944  [main] INFO  o.a.s.f.p.FluxParser - Not performing environment variable substitution.

Configuration (interpreted): 

!!org.apache.storm.flux.model.TopologyDef

bolts: []

components: []

config:

  es.indexer.addresses: localhost

  es.indexer.index.name: content

  es.indexer.create: false

  es.indexer.bulkActions: 100

  es.indexer.flushInterval: 2s

  es.indexer.concurrentRequests: 1

  es.indexer.settings: {cluster.name: elasticsearch}

  es.metrics.addresses: http://localhost:9200

  es.metrics.index.name: metrics

  es.metrics.settings: {cluster.name: elasticsearch}

  es.status.addresses: http://localhost:9200

  es.status.index.name: status

  es.status.routing: true

  es.status.routing.fieldname: key

  es.status.bulkActions: 500

  es.status.flushInterval: 5s

  es.status.concurrentRequests: 1

  es.status.settings: {cluster.name: elasticsearch}

  spout.ttl.purgatory: 30

  spout.min.delay.queries: 2000

  spout.reset.fetchdate.after: 120

  es.status.max.buckets: 50

  es.status.max.urls.per.bucket: 2

  es.status.bucket.field: key

  es.status.bucket.sort.field: [nextFetchDate, url]

  es.status.global.sort.field: nextFetchDate

  es.status.max.start.offset: 500

  es.status.sample: false

  es.status.query.timeout: -1

  es.status.recentDate.increase: -1

  es.status.recentDate.min.gap: -1

  topology.metrics.consumer.register:

  - {class: com.digitalpebble.stormcrawler.elasticsearch.metrics.MetricsConsumer,

    parallelism.hint: 1}

includes: null

name: null

spouts: []

streams: []

topologySource: null


955  [main] INFO  o.a.s.f.FluxBuilder - Detected DSL topology...

1399 [main] WARN  o.a.s.u.Utils - STORM-VERSION new 1.2.3 old null

---------- TOPOLOGY DETAILS ----------

Topology Name: crawler

--------------- SPOUTS ---------------

spout [10] (com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout)

---------------- BOLTS ---------------

partitioner [1] (com.digitalpebble.stormcrawler.bolt.URLPartitionerBolt)

fetcher [1] (com.digitalpebble.stormcrawler.bolt.FetcherBolt)

sitemap [1] (com.digitalpebble.stormcrawler.bolt.SiteMapParserBolt)

parse [1] (com.digitalpebble.stormcrawler.bolt.JSoupParserBolt)

index [1] (com.digitalpebble.stormcrawler.elasticsearch.bolt.IndexerBolt)

status [1] (com.digitalpebble.stormcrawler.elasticsearch.persistence.StatusUpdaterBolt)

status_metrics [1] (com.digitalpebble.stormcrawler.elasticsearch.metrics.StatusMetricsBolt)

--------------- STREAMS ---------------

spout --SHUFFLE--> partitioner

spout --SHUFFLE--> status_metrics

partitioner --FIELDS--> fetcher

fetcher --LOCAL_OR_SHUFFLE--> sitemap

sitemap --LOCAL_OR_SHUFFLE--> parse

parse --LOCAL_OR_SHUFFLE--> index

fetcher --FIELDS--> status

sitemap --FIELDS--> status

parse --FIELDS--> status

index --FIELDS--> status

--------------------------------------

1406 [main] INFO  o.a.s.f.Flux - Running in local mode...

4037 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:zookeeper.version=3.4.14-4c25d480e66aadd371de8bd2fd8da255ac140bcf, built on 03/06/2019 16:18 GMT

4038 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:host.name=ip-172-31-22-101.us-west-2.compute.internal

4038 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:java.version=11.0.4

4038 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:java.vendor=Ubuntu

4038 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:java.home=/usr/lib/jvm/java-11-openjdk-amd64

4039 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:java.class.path=/opt/apache-storm-1.2.3/*:/opt/apache-storm-1.2.3/lib/storm-rename-hack-1.2.3.jar:/opt/apache-storm-1.2.3/lib/disruptor-3.3.11.jar:/opt/apache-storm-1.2.3/lib/servlet-api-2.5.jar:/opt/apache-storm-1.2.3/lib/metrics-graphite-3.1.0.jar:/opt/apache-storm-1.2.3/lib/clojure-1.7.0.jar:/opt/apache-storm-1.2.3/lib/objenesis-2.1.jar:/opt/apache-storm-1.2.3/lib/asm-5.0.3.jar:/opt/apache-storm-1.2.3/lib/kryo-3.0.3.jar:/opt/apache-storm-1.2.3/lib/log4j-core-2.8.2.jar:/opt/apache-storm-1.2.3/lib/slf4j-api-1.7.21.jar:/opt/apache-storm-1.2.3/lib/minlog-1.3.0.jar:/opt/apache-storm-1.2.3/lib/reflectasm-1.10.1.jar:/opt/apache-storm-1.2.3/lib/log4j-over-slf4j-1.6.6.jar:/opt/apache-storm-1.2.3/lib/ring-cors-0.1.5.jar:/opt/apache-storm-1.2.3/lib/storm-core-1.2.3.jar:/opt/apache-storm-1.2.3/lib/metrics-core-3.1.0.jar:/opt/apache-storm-1.2.3/lib/log4j-api-2.8.2.jar:/opt/apache-storm-1.2.3/lib/log4j-slf4j-impl-2.8.2.jar:/opt/apache-storm-1.2.3/extlib/*:target/dev1-0.1.jar:/opt/apache-storm-1.2.3/conf:/opt/apache-storm-1.2.3/bin

4039 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:java.library.path=/usr/local/lib:/opt/local/lib:/usr/lib

4039 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:java.io.tmpdir=/tmp

4040 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:java.compiler=<NA>

4040 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:os.name=Linux

4040 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:os.arch=amd64

4040 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:os.version=4.15.0-1054-aws

4041 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:user.name=ubuntu

4041 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:user.home=/home/ubuntu

4041 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Server environment:user.dir=/home/ubuntu/crawl/dev1

4052 [main] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Created server with tickTime 2000 minSessionTimeout 4000 maxSessionTimeout 40000 datadir /tmp/bb22e646-cf52-47c9-8269-f32dc8116be1/version-2 snapdir /tmp/bb22e646-cf52-47c9-8269-f32dc8116be1/version-2

4068 [main] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - binding to port 0.0.0.0/0.0.0.0:2000

4070 [main] INFO  o.a.s.zookeeper - Starting inprocess zookeeper at port 2000 and dir /tmp/bb22e646-cf52-47c9-8269-f32dc8116be1

4158 [main] INFO  o.a.s.d.nimbus - Starting Nimbus with conf {topology.builtin.metrics.bucket.size.secs=60, nimbus.childopts=-Xmx1024m, ui.filter.params=null, storm.cluster.mode=local, storm.messaging.netty.client_worker_threads=1, logviewer.max.per.worker.logs.size.mb=2048, supervisor.run.worker.as.user=false, topology.max.task.parallelism=null, topology.priority=29, zmq.threads=1, storm.group.mapping.service=org.apache.storm.security.auth.ShellBasedGroupsMapping, transactional.zookeeper.root=/transactional, topology.sleep.spout.wait.strategy.time.ms=1, ui.pagination=20, scheduler.display.resource=false, topology.max.replication.wait.time.sec=60, drpc.invocations.port=3773, supervisor.localizer.cache.target.size.mb=10240, topology.multilang.serializer=org.apache.storm.multilang.JsonSerializer, storm.messaging.netty.server_worker_threads=1, nimbus.blobstore.class=org.apache.storm.blobstore.LocalFsBlobStore, resource.aware.scheduler.eviction.strategy=org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy, topology.max.error.report.per.interval=5, storm.thrift.transport=org.apache.storm.security.auth.SimpleTransportPlugin, zmq.hwm=0, storm.group.mapping.service.params=null, worker.profiler.enabled=false, storm.principal.tolocal=org.apache.storm.security.auth.DefaultPrincipalToLocal, supervisor.worker.shutdown.sleep.secs=3, pacemaker.host=localhost, storm.zookeeper.retry.times=5, ui.actions.enabled=true, zmq.linger.millis=0, supervisor.enable=true, topology.stats.sample.rate=0.05, storm.nimbus.zookeeper.acls.fixup=true, storm.messaging.netty.min_wait_ms=100, worker.log.level.reset.poll.secs=30, storm.zookeeper.port=2000, supervisor.heartbeat.frequency.secs=5, topology.enable.message.timeouts=true, supervisor.cpu.capacity=400.0, drpc.worker.threads=64, supervisor.blobstore.download.thread.count=5, task.backpressure.poll.secs=30, drpc.queue.size=128, topology.backpressure.enable=false, supervisor.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.blobstore.inputstream.buffer.size.bytes=65536, topology.shellbolt.max.pending=100, drpc.https.keystore.password=*****, nimbus.code.sync.freq.secs=120, logviewer.port=8000, topology.scheduler.strategy=org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy, topology.executor.send.buffer.size=1024, resource.aware.scheduler.priority.strategy=org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy, pacemaker.auth.method=NONE, storm.daemon.metrics.reporter.plugins=["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], topology.worker.logwriter.childopts=-Xmx64m, topology.spout.wait.strategy=org.apache.storm.spout.SleepSpoutWaitStrategy, ui.host=0.0.0.0, storm.nimbus.retry.interval.millis=2000, nimbus.inbox.jar.expiration.secs=3600, dev.zookeeper.path=/tmp/dev-storm-zookeeper, topology.acker.executors=null, topology.fall.back.on.java.serialization=true, topology.eventlogger.executors=0, supervisor.localizer.cleanup.interval.ms=600000, storm.zookeeper.servers=["localhost"], nimbus.thrift.threads=64, logviewer.cleanup.age.mins=10080, topology.worker.childopts=null, topology.classpath=null, supervisor.monitor.frequency.secs=3, nimbus.credential.renewers.freq.secs=600, backpressure.znode.timeout.secs=30, topology.skip.missing.kryo.registrations=true, drpc.authorizer.acl.filename=drpc-auth-acl.yaml, pacemaker.kerberos.users=[], storm.group.mapping.service.cache.duration.secs=120, blobstore.dir=/tmp/55b9f681-e0ff-4929-81ff-d3354e5f786e, topology.testing.always.try.serialize=false, nimbus.monitor.freq.secs=10, storm.health.check.timeout.ms=5000, supervisor.supervisors=[], topology.tasks=null, topology.bolts.outgoing.overflow.buffer.enable=false, storm.messaging.netty.socket.backlog=500, topology.workers=1, pacemaker.base.threads=10, storm.local.dir=/tmp/55b9f681-e0ff-4929-81ff-d3354e5f786e, backpressure.znode.update.freq.secs=15, worker.childopts=-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump, storm.auth.simple-white-list.users=[], topology.disruptor.batch.timeout.millis=1, topology.message.timeout.secs=30, topology.state.synchronization.timeout.secs=60, topology.tuple.serializer=org.apache.storm.serialization.types.ListDelegateSerializer, supervisor.supervisors.commands=[], nimbus.blobstore.expiration.secs=600, logviewer.childopts=-Xmx128m, storm.nimbus.zookeeper.acls.check=true, topology.environment=null, topology.debug=false, topology.disruptor.batch.size=100, storm.disable.symlinks=false, storm.messaging.netty.max_retries=300, ui.childopts=-Xmx768m, storm.network.topography.plugin=org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping, storm.zookeeper.session.timeout=20000, drpc.childopts=-Xmx768m, drpc.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.connection.timeout=15000, storm.zookeeper.auth.user=null, storm.meta.serialization.delegate=org.apache.storm.serialization.GzipThriftSerializationDelegate, topology.max.spout.pending=null, storm.codedistributor.class=org.apache.storm.codedistributor.LocalFileSystemCodeDistributor, nimbus.supervisor.timeout.secs=60, nimbus.task.timeout.secs=30, drpc.port=3772, pacemaker.max.threads=50, storm.zookeeper.retry.intervalceiling.millis=30000, nimbus.thrift.port=6627, storm.auth.simple-acl.admins=[], topology.component.cpu.pcore.percent=10.0, supervisor.memory.capacity.mb=3072.0, storm.nimbus.retry.times=5, supervisor.worker.start.timeout.secs=120, storm.zookeeper.retry.interval=1000, logs.users=null, storm.cluster.metrics.consumer.publish.interval.secs=60, worker.profiler.command=flight.bash, transactional.zookeeper.port=null, drpc.max_buffer_size=1048576, pacemaker.thread.timeout=10, task.credentials.poll.secs=30, blobstore.superuser=ubuntu, drpc.https.keystore.type=JKS, topology.worker.receiver.thread.count=1, topology.state.checkpoint.interval.ms=1000, supervisor.slots.ports=[6700 6701 6702 6703], topology.transfer.buffer.size=1024, storm.health.check.dir=healthchecks, topology.worker.shared.thread.pool.size=4, drpc.authorizer.acl.strict=false, nimbus.file.copy.expiration.secs=600, worker.profiler.childopts=-XX:+UnlockCommercialFeatures -XX:+FlightRecorder, topology.executor.receive.buffer.size=1024, backpressure.disruptor.low.watermark=0.4, nimbus.task.launch.secs=120, storm.local.mode.zmq=false, storm.messaging.netty.buffer_size=5242880, storm.cluster.state.store=org.apache.storm.cluster_state.zookeeper_state_factory, worker.heartbeat.frequency.secs=1, storm.log4j2.conf.dir=log4j2, ui.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.root=/storm, topology.tick.tuple.freq.secs=null, drpc.https.port=-1, storm.workers.artifacts.dir=workers-artifacts, supervisor.blobstore.download.max_retries=3, task.refresh.poll.secs=10, storm.exhibitor.port=8080, task.heartbeat.frequency.secs=3, pacemaker.port=6699, storm.messaging.netty.max_wait_ms=1000, topology.component.resources.offheap.memory.mb=0.0, drpc.http.port=3774, topology.error.throttle.interval.secs=10, storm.messaging.transport=org.apache.storm.messaging.netty.Context, topology.disable.loadaware.messaging=false, storm.messaging.netty.authentication=false, topology.component.resources.onheap.memory.mb=128.0, topology.kryo.factory=org.apache.storm.serialization.DefaultKryoFactory, worker.gc.childopts=, nimbus.topology.validator=org.apache.storm.nimbus.DefaultTopologyValidator, nimbus.seeds=["localhost"], nimbus.queue.size=100000, nimbus.cleanup.inbox.freq.secs=600, storm.blobstore.replication.factor=3, worker.heap.memory.mb=768, logviewer.max.sum.worker.logs.size.mb=4096, pacemaker.childopts=-Xmx1024m, ui.users=null, transactional.zookeeper.servers=null, supervisor.worker.timeout.secs=30, storm.zookeeper.auth.password=null, storm.blobstore.acl.validation.enabled=false, client.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.thrift.socket.timeout.ms=600000, supervisor.childopts=-Xmx256m, topology.worker.max.heap.size.mb=768.0, ui.http.x-frame-options=DENY, backpressure.disruptor.high.watermark=0.9, ui.filter=null, ui.header.buffer.bytes=4096, topology.min.replication.count=1, topology.disruptor.wait.timeout.millis=1000, storm.nimbus.retry.intervalceiling.millis=60000, topology.trident.batch.emit.interval.millis=50, storm.auth.simple-acl.users=[], drpc.invocations.threads=64, java.library.path=/usr/local/lib:/opt/local/lib:/usr/lib, ui.port=8080, storm.exhibitor.poll.uripath=/exhibitor/v1/cluster/list, storm.messaging.netty.transfer.batch.size=262144, logviewer.appender.name=A1, nimbus.thrift.max_buffer_size=1048576, storm.auth.simple-acl.users.commands=[], drpc.request.timeout.secs=600}

4165 [main] INFO  o.a.s.n.NimbusInfo - Nimbus figures out its name to ip-172-31-22-101.us-west-2.compute.internal

4212 [main] INFO  o.a.s.s.o.a.c.u.Compatibility - Running in ZooKeeper 3.4.x compatibility mode

4238 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4249 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:zookeeper.version=3.4.14-4c25d480e66aadd371de8bd2fd8da255ac140bcf, built on 03/06/2019 16:18 GMT

4250 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:host.name=ip-172-31-22-101.us-west-2.compute.internal

4250 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:java.version=11.0.4

4250 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:java.vendor=Ubuntu

4250 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:java.home=/usr/lib/jvm/java-11-openjdk-amd64

4251 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:java.class.path=/opt/apache-storm-1.2.3/*:/opt/apache-storm-1.2.3/lib/storm-rename-hack-1.2.3.jar:/opt/apache-storm-1.2.3/lib/disruptor-3.3.11.jar:/opt/apache-storm-1.2.3/lib/servlet-api-2.5.jar:/opt/apache-storm-1.2.3/lib/metrics-graphite-3.1.0.jar:/opt/apache-storm-1.2.3/lib/clojure-1.7.0.jar:/opt/apache-storm-1.2.3/lib/objenesis-2.1.jar:/opt/apache-storm-1.2.3/lib/asm-5.0.3.jar:/opt/apache-storm-1.2.3/lib/kryo-3.0.3.jar:/opt/apache-storm-1.2.3/lib/log4j-core-2.8.2.jar:/opt/apache-storm-1.2.3/lib/slf4j-api-1.7.21.jar:/opt/apache-storm-1.2.3/lib/minlog-1.3.0.jar:/opt/apache-storm-1.2.3/lib/reflectasm-1.10.1.jar:/opt/apache-storm-1.2.3/lib/log4j-over-slf4j-1.6.6.jar:/opt/apache-storm-1.2.3/lib/ring-cors-0.1.5.jar:/opt/apache-storm-1.2.3/lib/storm-core-1.2.3.jar:/opt/apache-storm-1.2.3/lib/metrics-core-3.1.0.jar:/opt/apache-storm-1.2.3/lib/log4j-api-2.8.2.jar:/opt/apache-storm-1.2.3/lib/log4j-slf4j-impl-2.8.2.jar:/opt/apache-storm-1.2.3/extlib/*:target/dev1-0.1.jar:/opt/apache-storm-1.2.3/conf:/opt/apache-storm-1.2.3/bin

4251 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:java.library.path=/usr/local/lib:/opt/local/lib:/usr/lib

4251 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:java.io.tmpdir=/tmp

4251 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:java.compiler=<NA>

4252 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:os.name=Linux

4252 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:os.arch=amd64

4253 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:os.version=4.15.0-1054-aws

4253 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:user.name=ubuntu

4253 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:user.home=/home/ubuntu

4254 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Client environment:user.dir=/home/ubuntu/crawl/dev1

4255 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@796eeafb

4267 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4275 [main] INFO  o.a.s.b.FileBlobStoreImpl - Creating new blob store based in /tmp/55b9f681-e0ff-4929-81ff-d3354e5f786e/blobs

4281 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4287 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@196519dd

4289 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4289 [main] INFO  o.a.s.d.nimbus - Using default scheduler

4289 [main] INFO  o.a.s.n.NimbusInfo - Nimbus figures out its name to ip-172-31-22-101.us-west-2.compute.internal

4309 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4311 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4311 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55310

4312 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4311 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4318 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55312

4320 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55312

4322 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.p.FileTxnLog - Creating new log file: log.1

4322 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55310

4345 [main] INFO  o.a.s.n.NimbusInfo - Nimbus figures out its name to ip-172-31-22-101.us-west-2.compute.internal

4347 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f20000, negotiated timeout = 20000

4347 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f20000 with negotiated timeout 20000 for client /127.0.0.1:55312

4358 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4359 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f20001 with negotiated timeout 20000 for client /127.0.0.1:55310

4359 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f20001, negotiated timeout = 20000

4360 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4362 [main-EventThread] INFO  o.a.s.zookeeper - Zookeeper state update: :connected:none

4374 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4375 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@2778f3d1

4376 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4382 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4383 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4383 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55314

4383 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55314

4384 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f20002 with negotiated timeout 20000 for client /127.0.0.1:55314

4385 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f20002, negotiated timeout = 20000

4385 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4386 [main-EventThread] INFO  o.a.s.zookeeper - Zookeeper state update: :connected:none

4410 [Curator-Framework-0] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting

4412 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x100040091f20002

4414 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:55314 which had sessionid 0x100040091f20002

4415 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Session: 0x100040091f20002 closed

4415 [main-EventThread] INFO  o.a.s.s.o.a.z.ClientCnxn - EventThread shut down for session: 0x100040091f20002

4417 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4418 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@40c858fc

4423 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4425 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4426 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4426 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55316

4427 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4427 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55316

4428 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@6cf7de4e

4432 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4435 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4436 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4434 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f20003, negotiated timeout = 20000

4436 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4434 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f20003 with negotiated timeout 20000 for client /127.0.0.1:55316

4436 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55318

4437 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55318

4440 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f20004 with negotiated timeout 20000 for client /127.0.0.1:55318

4440 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f20004, negotiated timeout = 20000

4442 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4490 [main] INFO  o.a.s.zookeeper - Queued up for leader lock.

4506 [main] INFO  o.a.s.d.m.MetricsUtils - Using statistics reporter plugin:org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter

4511 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f20000 type:create cxid:0x1 zxid:0x12 txntype:-1 reqpath:n/a Error Path:/storm/leader-lock Error:KeeperErrorCode = NoNode for /storm/leader-lock

4512 [main] INFO  o.a.s.d.m.r.JmxPreparableReporter - Preparing...

4514 [Curator-Framework-0] WARN  o.a.s.s.o.a.c.u.ZKPaths - The version of ZooKeeper being used doesn't support Container nodes. CreateMode.PERSISTENT will be used instead.

4532 [main-EventThread] INFO  o.a.s.z.Zookeeper - active-topology-blobs [] local-topology-blobs [] diff-topology-blobs []

4533 [main-EventThread] INFO  o.a.s.z.Zookeeper - active-topology-dependencies [] local-blobs [] diff-topology-dependencies []

4533 [main-EventThread] INFO  o.a.s.z.Zookeeper - Accepting leadership, all active topologies and corresponding dependencies found locally.

4543 [main] INFO  o.a.s.d.common - Started statistics report plugin...

4557 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4557 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@74de1e9d

4562 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4564 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4564 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55320

4564 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55320

4564 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4566 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f20005 with negotiated timeout 20000 for client /127.0.0.1:55320

4566 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f20005, negotiated timeout = 20000

4566 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4567 [main-EventThread] INFO  o.a.s.zookeeper - Zookeeper state update: :connected:none

4569 [Curator-Framework-0] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting

4569 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x100040091f20005

4571 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:55320 which had sessionid 0x100040091f20005

4571 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Session: 0x100040091f20005 closed

4572 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4571 [main-EventThread] INFO  o.a.s.s.o.a.z.ClientCnxn - EventThread shut down for session: 0x100040091f20005

4578 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@191d69a

4579 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4579 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4580 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@23043ba

4580 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4585 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4585 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4585 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55322

4585 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4586 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4586 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55324

4586 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55322

4586 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55324

4588 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f20006 with negotiated timeout 20000 for client /127.0.0.1:55322

4588 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f20006, negotiated timeout = 20000

4589 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4590 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f20007 with negotiated timeout 20000 for client /127.0.0.1:55324

4590 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f20007, negotiated timeout = 20000

4590 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4591 [main-EventThread] INFO  o.a.s.zookeeper - Zookeeper state update: :connected:none

4592 [Curator-Framework-0] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting

4593 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x100040091f20007

4594 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:55324 which had sessionid 0x100040091f20007

4594 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Session: 0x100040091f20007 closed

4595 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4599 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@1b47e09c

4599 [main-EventThread] INFO  o.a.s.s.o.a.z.ClientCnxn - EventThread shut down for session: 0x100040091f20007

4611 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4611 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55326

4611 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4612 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55326

4612 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4613 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f20008 with negotiated timeout 20000 for client /127.0.0.1:55326

4614 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f20008, negotiated timeout = 20000

4614 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4642 [main] INFO  o.a.s.z.Zookeeper - Staring ZK Curator

4642 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4643 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@188ccb6c

4648 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4651 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4652 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4652 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55328

4653 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55328

4654 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f20009 with negotiated timeout 20000 for client /127.0.0.1:55328

4655 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f20009, negotiated timeout = 20000

4655 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4658 [Curator-Framework-0] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting

4659 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x100040091f20009

4660 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Session: 0x100040091f20009 closed

4660 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:55328 which had sessionid 0x100040091f20009

4663 [main] INFO  o.a.s.z.Zookeeper - Staring ZK Curator

4663 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4660 [main-EventThread] INFO  o.a.s.s.o.a.z.ClientCnxn - EventThread shut down for session: 0x100040091f20009

4665 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@79753f20

4670 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4671 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4672 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4672 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55330

4673 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55330

4674 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f2000a with negotiated timeout 20000 for client /127.0.0.1:55330

4675 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f2000a, negotiated timeout = 20000

4677 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4686 [main] INFO  o.a.s.l.Localizer - Reconstruct localized resource: /tmp/ddf53adf-8313-4830-8786-4bd66e78b4fe/supervisor/usercache

4686 [main] WARN  o.a.s.l.Localizer - No left over resources found for any user during reconstructing of local resources at: /tmp/ddf53adf-8313-4830-8786-4bd66e78b4fe/supervisor/usercache

4691 [main] INFO  o.a.s.d.s.Supervisor - Starting Supervisor with conf {topology.builtin.metrics.bucket.size.secs=60, nimbus.childopts=-Xmx1024m, ui.filter.params=null, storm.cluster.mode=local, storm.messaging.netty.client_worker_threads=1, logviewer.max.per.worker.logs.size.mb=2048, supervisor.run.worker.as.user=false, topology.max.task.parallelism=null, topology.priority=29, zmq.threads=1, storm.group.mapping.service=org.apache.storm.security.auth.ShellBasedGroupsMapping, transactional.zookeeper.root=/transactional, topology.sleep.spout.wait.strategy.time.ms=1, ui.pagination=20, scheduler.display.resource=false, topology.max.replication.wait.time.sec=60, drpc.invocations.port=3773, supervisor.localizer.cache.target.size.mb=10240, topology.multilang.serializer=org.apache.storm.multilang.JsonSerializer, storm.messaging.netty.server_worker_threads=1, nimbus.blobstore.class=org.apache.storm.blobstore.LocalFsBlobStore, resource.aware.scheduler.eviction.strategy=org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy, topology.max.error.report.per.interval=5, storm.thrift.transport=org.apache.storm.security.auth.SimpleTransportPlugin, zmq.hwm=0, storm.group.mapping.service.params=null, worker.profiler.enabled=false, storm.principal.tolocal=org.apache.storm.security.auth.DefaultPrincipalToLocal, supervisor.worker.shutdown.sleep.secs=3, pacemaker.host=localhost, storm.zookeeper.retry.times=5, ui.actions.enabled=true, zmq.linger.millis=0, supervisor.enable=true, topology.stats.sample.rate=0.05, storm.nimbus.zookeeper.acls.fixup=true, storm.messaging.netty.min_wait_ms=100, worker.log.level.reset.poll.secs=30, storm.zookeeper.port=2000, supervisor.heartbeat.frequency.secs=5, topology.enable.message.timeouts=true, supervisor.cpu.capacity=400.0, drpc.worker.threads=64, supervisor.blobstore.download.thread.count=5, task.backpressure.poll.secs=30, drpc.queue.size=128, topology.backpressure.enable=false, supervisor.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.blobstore.inputstream.buffer.size.bytes=65536, topology.shellbolt.max.pending=100, drpc.https.keystore.password=*****, nimbus.code.sync.freq.secs=120, logviewer.port=8000, topology.scheduler.strategy=org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy, topology.executor.send.buffer.size=1024, resource.aware.scheduler.priority.strategy=org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy, pacemaker.auth.method=NONE, storm.daemon.metrics.reporter.plugins=[org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter], topology.worker.logwriter.childopts=-Xmx64m, topology.spout.wait.strategy=org.apache.storm.spout.SleepSpoutWaitStrategy, ui.host=0.0.0.0, storm.nimbus.retry.interval.millis=2000, nimbus.inbox.jar.expiration.secs=3600, dev.zookeeper.path=/tmp/dev-storm-zookeeper, topology.acker.executors=null, topology.fall.back.on.java.serialization=true, topology.eventlogger.executors=0, supervisor.localizer.cleanup.interval.ms=600000, storm.zookeeper.servers=[localhost], nimbus.thrift.threads=64, logviewer.cleanup.age.mins=10080, topology.worker.childopts=null, topology.classpath=null, supervisor.monitor.frequency.secs=3, nimbus.credential.renewers.freq.secs=600, backpressure.znode.timeout.secs=30, topology.skip.missing.kryo.registrations=true, drpc.authorizer.acl.filename=drpc-auth-acl.yaml, pacemaker.kerberos.users=[], storm.group.mapping.service.cache.duration.secs=120, blobstore.dir=/tmp/55b9f681-e0ff-4929-81ff-d3354e5f786e, topology.testing.always.try.serialize=false, nimbus.monitor.freq.secs=10, storm.health.check.timeout.ms=5000, supervisor.supervisors=[], topology.tasks=null, topology.bolts.outgoing.overflow.buffer.enable=false, storm.messaging.netty.socket.backlog=500, topology.workers=1, pacemaker.base.threads=10, storm.local.dir=/tmp/ddf53adf-8313-4830-8786-4bd66e78b4fe, backpressure.znode.update.freq.secs=15, worker.childopts=-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump, storm.auth.simple-white-list.users=[], topology.disruptor.batch.timeout.millis=1, topology.message.timeout.secs=30, topology.state.synchronization.timeout.secs=60, topology.tuple.serializer=org.apache.storm.serialization.types.ListDelegateSerializer, supervisor.supervisors.commands=[], nimbus.blobstore.expiration.secs=600, logviewer.childopts=-Xmx128m, storm.nimbus.zookeeper.acls.check=true, topology.environment=null, topology.debug=false, topology.disruptor.batch.size=100, storm.disable.symlinks=false, storm.messaging.netty.max_retries=300, ui.childopts=-Xmx768m, storm.network.topography.plugin=org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping, storm.zookeeper.session.timeout=20000, drpc.childopts=-Xmx768m, drpc.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.connection.timeout=15000, storm.zookeeper.auth.user=null, storm.meta.serialization.delegate=org.apache.storm.serialization.GzipThriftSerializationDelegate, topology.max.spout.pending=null, storm.codedistributor.class=org.apache.storm.codedistributor.LocalFileSystemCodeDistributor, nimbus.supervisor.timeout.secs=60, nimbus.task.timeout.secs=30, drpc.port=3772, pacemaker.max.threads=50, storm.zookeeper.retry.intervalceiling.millis=30000, nimbus.thrift.port=6627, storm.auth.simple-acl.admins=[], topology.component.cpu.pcore.percent=10.0, supervisor.memory.capacity.mb=3072.0, storm.nimbus.retry.times=5, supervisor.worker.start.timeout.secs=120, storm.zookeeper.retry.interval=1000, logs.users=null, storm.cluster.metrics.consumer.publish.interval.secs=60, worker.profiler.command=flight.bash, transactional.zookeeper.port=null, drpc.max_buffer_size=1048576, pacemaker.thread.timeout=10, task.credentials.poll.secs=30, blobstore.superuser=ubuntu, drpc.https.keystore.type=JKS, topology.worker.receiver.thread.count=1, topology.state.checkpoint.interval.ms=1000, supervisor.slots.ports=[1024, 1025, 1026], topology.transfer.buffer.size=1024, storm.health.check.dir=healthchecks, topology.worker.shared.thread.pool.size=4, drpc.authorizer.acl.strict=false, nimbus.file.copy.expiration.secs=600, worker.profiler.childopts=-XX:+UnlockCommercialFeatures -XX:+FlightRecorder, topology.executor.receive.buffer.size=1024, backpressure.disruptor.low.watermark=0.4, nimbus.task.launch.secs=120, storm.local.mode.zmq=false, storm.messaging.netty.buffer_size=5242880, storm.cluster.state.store=org.apache.storm.cluster_state.zookeeper_state_factory, worker.heartbeat.frequency.secs=1, storm.log4j2.conf.dir=log4j2, ui.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.root=/storm, topology.tick.tuple.freq.secs=null, drpc.https.port=-1, storm.workers.artifacts.dir=workers-artifacts, supervisor.blobstore.download.max_retries=3, task.refresh.poll.secs=10, storm.exhibitor.port=8080, task.heartbeat.frequency.secs=3, pacemaker.port=6699, storm.messaging.netty.max_wait_ms=1000, topology.component.resources.offheap.memory.mb=0.0, drpc.http.port=3774, topology.error.throttle.interval.secs=10, storm.messaging.transport=org.apache.storm.messaging.netty.Context, topology.disable.loadaware.messaging=false, storm.messaging.netty.authentication=false, topology.component.resources.onheap.memory.mb=128.0, topology.kryo.factory=org.apache.storm.serialization.DefaultKryoFactory, worker.gc.childopts=, nimbus.topology.validator=org.apache.storm.nimbus.DefaultTopologyValidator, nimbus.seeds=[localhost], nimbus.queue.size=100000, nimbus.cleanup.inbox.freq.secs=600, storm.blobstore.replication.factor=3, worker.heap.memory.mb=768, logviewer.max.sum.worker.logs.size.mb=4096, pacemaker.childopts=-Xmx1024m, ui.users=null, transactional.zookeeper.servers=null, supervisor.worker.timeout.secs=30, storm.zookeeper.auth.password=null, storm.blobstore.acl.validation.enabled=false, client.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.thrift.socket.timeout.ms=600000, supervisor.childopts=-Xmx256m, topology.worker.max.heap.size.mb=768.0, ui.http.x-frame-options=DENY, backpressure.disruptor.high.watermark=0.9, ui.filter=null, ui.header.buffer.bytes=4096, topology.min.replication.count=1, topology.disruptor.wait.timeout.millis=1000, storm.nimbus.retry.intervalceiling.millis=60000, topology.trident.batch.emit.interval.millis=50, storm.auth.simple-acl.users=[], drpc.invocations.threads=64, java.library.path=/usr/local/lib:/opt/local/lib:/usr/lib, ui.port=8080, storm.exhibitor.poll.uripath=/exhibitor/v1/cluster/list, storm.messaging.netty.transfer.batch.size=262144, logviewer.appender.name=A1, nimbus.thrift.max_buffer_size=1048576, storm.auth.simple-acl.users.commands=[], drpc.request.timeout.secs=600}

4704 [main] WARN  o.a.s.d.s.Slot - SLOT ip-172-31-22-101.us-west-2.compute.internal:1024 Starting in state EMPTY - assignment null

4704 [main] WARN  o.a.s.d.s.Slot - SLOT ip-172-31-22-101.us-west-2.compute.internal:1025 Starting in state EMPTY - assignment null

4705 [main] WARN  o.a.s.d.s.Slot - SLOT ip-172-31-22-101.us-west-2.compute.internal:1026 Starting in state EMPTY - assignment null

4705 [main] INFO  o.a.s.l.AsyncLocalizer - Cleaning up unused topologies in /tmp/ddf53adf-8313-4830-8786-4bd66e78b4fe/supervisor/stormdist

4707 [main] INFO  o.a.s.d.s.Supervisor - Starting supervisor with id ab823f98-11b3-4021-b8e7-f9da7ebc12f3 at host ip-172-31-22-101.us-west-2.compute.internal.

4708 [main] INFO  o.a.s.z.Zookeeper - Staring ZK Curator

4709 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4709 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@405a296e

4711 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4712 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4712 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4712 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55332

4712 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55332

4714 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f2000b with negotiated timeout 20000 for client /127.0.0.1:55332

4714 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f2000b, negotiated timeout = 20000

4715 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4718 [Curator-Framework-0] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting

4718 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x100040091f2000b

4719 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:55332 which had sessionid 0x100040091f2000b

4719 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Session: 0x100040091f2000b closed

4720 [main-EventThread] INFO  o.a.s.s.o.a.z.ClientCnxn - EventThread shut down for session: 0x100040091f2000b

4721 [main] INFO  o.a.s.z.Zookeeper - Staring ZK Curator

4721 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

4723 [main] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@12f85dc8

4732 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

4732 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

4732 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55334

4732 [main] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

4732 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55334

4734 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f2000c with negotiated timeout 20000 for client /127.0.0.1:55334

4734 [main-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f2000c, negotiated timeout = 20000

4735 [main-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

4741 [main] INFO  o.a.s.l.Localizer - Reconstruct localized resource: /tmp/a84faee2-1f19-46b6-affa-3c9d4bb2d61a/supervisor/usercache

4741 [main] WARN  o.a.s.l.Localizer - No left over resources found for any user during reconstructing of local resources at: /tmp/a84faee2-1f19-46b6-affa-3c9d4bb2d61a/supervisor/usercache

4743 [main] INFO  o.a.s.d.s.Supervisor - Starting Supervisor with conf {topology.builtin.metrics.bucket.size.secs=60, nimbus.childopts=-Xmx1024m, ui.filter.params=null, storm.cluster.mode=local, storm.messaging.netty.client_worker_threads=1, logviewer.max.per.worker.logs.size.mb=2048, supervisor.run.worker.as.user=false, topology.max.task.parallelism=null, topology.priority=29, zmq.threads=1, storm.group.mapping.service=org.apache.storm.security.auth.ShellBasedGroupsMapping, transactional.zookeeper.root=/transactional, topology.sleep.spout.wait.strategy.time.ms=1, ui.pagination=20, scheduler.display.resource=false, topology.max.replication.wait.time.sec=60, drpc.invocations.port=3773, supervisor.localizer.cache.target.size.mb=10240, topology.multilang.serializer=org.apache.storm.multilang.JsonSerializer, storm.messaging.netty.server_worker_threads=1, nimbus.blobstore.class=org.apache.storm.blobstore.LocalFsBlobStore, resource.aware.scheduler.eviction.strategy=org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy, topology.max.error.report.per.interval=5, storm.thrift.transport=org.apache.storm.security.auth.SimpleTransportPlugin, zmq.hwm=0, storm.group.mapping.service.params=null, worker.profiler.enabled=false, storm.principal.tolocal=org.apache.storm.security.auth.DefaultPrincipalToLocal, supervisor.worker.shutdown.sleep.secs=3, pacemaker.host=localhost, storm.zookeeper.retry.times=5, ui.actions.enabled=true, zmq.linger.millis=0, supervisor.enable=true, topology.stats.sample.rate=0.05, storm.nimbus.zookeeper.acls.fixup=true, storm.messaging.netty.min_wait_ms=100, worker.log.level.reset.poll.secs=30, storm.zookeeper.port=2000, supervisor.heartbeat.frequency.secs=5, topology.enable.message.timeouts=true, supervisor.cpu.capacity=400.0, drpc.worker.threads=64, supervisor.blobstore.download.thread.count=5, task.backpressure.poll.secs=30, drpc.queue.size=128, topology.backpressure.enable=false, supervisor.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.blobstore.inputstream.buffer.size.bytes=65536, topology.shellbolt.max.pending=100, drpc.https.keystore.password=*****, nimbus.code.sync.freq.secs=120, logviewer.port=8000, topology.scheduler.strategy=org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy, topology.executor.send.buffer.size=1024, resource.aware.scheduler.priority.strategy=org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy, pacemaker.auth.method=NONE, storm.daemon.metrics.reporter.plugins=[org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter], topology.worker.logwriter.childopts=-Xmx64m, topology.spout.wait.strategy=org.apache.storm.spout.SleepSpoutWaitStrategy, ui.host=0.0.0.0, storm.nimbus.retry.interval.millis=2000, nimbus.inbox.jar.expiration.secs=3600, dev.zookeeper.path=/tmp/dev-storm-zookeeper, topology.acker.executors=null, topology.fall.back.on.java.serialization=true, topology.eventlogger.executors=0, supervisor.localizer.cleanup.interval.ms=600000, storm.zookeeper.servers=[localhost], nimbus.thrift.threads=64, logviewer.cleanup.age.mins=10080, topology.worker.childopts=null, topology.classpath=null, supervisor.monitor.frequency.secs=3, nimbus.credential.renewers.freq.secs=600, backpressure.znode.timeout.secs=30, topology.skip.missing.kryo.registrations=true, drpc.authorizer.acl.filename=drpc-auth-acl.yaml, pacemaker.kerberos.users=[], storm.group.mapping.service.cache.duration.secs=120, blobstore.dir=/tmp/55b9f681-e0ff-4929-81ff-d3354e5f786e, topology.testing.always.try.serialize=false, nimbus.monitor.freq.secs=10, storm.health.check.timeout.ms=5000, supervisor.supervisors=[], topology.tasks=null, topology.bolts.outgoing.overflow.buffer.enable=false, storm.messaging.netty.socket.backlog=500, topology.workers=1, pacemaker.base.threads=10, storm.local.dir=/tmp/a84faee2-1f19-46b6-affa-3c9d4bb2d61a, backpressure.znode.update.freq.secs=15, worker.childopts=-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump, storm.auth.simple-white-list.users=[], topology.disruptor.batch.timeout.millis=1, topology.message.timeout.secs=30, topology.state.synchronization.timeout.secs=60, topology.tuple.serializer=org.apache.storm.serialization.types.ListDelegateSerializer, supervisor.supervisors.commands=[], nimbus.blobstore.expiration.secs=600, logviewer.childopts=-Xmx128m, storm.nimbus.zookeeper.acls.check=true, topology.environment=null, topology.debug=false, topology.disruptor.batch.size=100, storm.disable.symlinks=false, storm.messaging.netty.max_retries=300, ui.childopts=-Xmx768m, storm.network.topography.plugin=org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping, storm.zookeeper.session.timeout=20000, drpc.childopts=-Xmx768m, drpc.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.connection.timeout=15000, storm.zookeeper.auth.user=null, storm.meta.serialization.delegate=org.apache.storm.serialization.GzipThriftSerializationDelegate, topology.max.spout.pending=null, storm.codedistributor.class=org.apache.storm.codedistributor.LocalFileSystemCodeDistributor, nimbus.supervisor.timeout.secs=60, nimbus.task.timeout.secs=30, drpc.port=3772, pacemaker.max.threads=50, storm.zookeeper.retry.intervalceiling.millis=30000, nimbus.thrift.port=6627, storm.auth.simple-acl.admins=[], topology.component.cpu.pcore.percent=10.0, supervisor.memory.capacity.mb=3072.0, storm.nimbus.retry.times=5, supervisor.worker.start.timeout.secs=120, storm.zookeeper.retry.interval=1000, logs.users=null, storm.cluster.metrics.consumer.publish.interval.secs=60, worker.profiler.command=flight.bash, transactional.zookeeper.port=null, drpc.max_buffer_size=1048576, pacemaker.thread.timeout=10, task.credentials.poll.secs=30, blobstore.superuser=ubuntu, drpc.https.keystore.type=JKS, topology.worker.receiver.thread.count=1, topology.state.checkpoint.interval.ms=1000, supervisor.slots.ports=[1027, 1028, 1029], topology.transfer.buffer.size=1024, storm.health.check.dir=healthchecks, topology.worker.shared.thread.pool.size=4, drpc.authorizer.acl.strict=false, nimbus.file.copy.expiration.secs=600, worker.profiler.childopts=-XX:+UnlockCommercialFeatures -XX:+FlightRecorder, topology.executor.receive.buffer.size=1024, backpressure.disruptor.low.watermark=0.4, nimbus.task.launch.secs=120, storm.local.mode.zmq=false, storm.messaging.netty.buffer_size=5242880, storm.cluster.state.store=org.apache.storm.cluster_state.zookeeper_state_factory, worker.heartbeat.frequency.secs=1, storm.log4j2.conf.dir=log4j2, ui.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.root=/storm, topology.tick.tuple.freq.secs=null, drpc.https.port=-1, storm.workers.artifacts.dir=workers-artifacts, supervisor.blobstore.download.max_retries=3, task.refresh.poll.secs=10, storm.exhibitor.port=8080, task.heartbeat.frequency.secs=3, pacemaker.port=6699, storm.messaging.netty.max_wait_ms=1000, topology.component.resources.offheap.memory.mb=0.0, drpc.http.port=3774, topology.error.throttle.interval.secs=10, storm.messaging.transport=org.apache.storm.messaging.netty.Context, topology.disable.loadaware.messaging=false, storm.messaging.netty.authentication=false, topology.component.resources.onheap.memory.mb=128.0, topology.kryo.factory=org.apache.storm.serialization.DefaultKryoFactory, worker.gc.childopts=, nimbus.topology.validator=org.apache.storm.nimbus.DefaultTopologyValidator, nimbus.seeds=[localhost], nimbus.queue.size=100000, nimbus.cleanup.inbox.freq.secs=600, storm.blobstore.replication.factor=3, worker.heap.memory.mb=768, logviewer.max.sum.worker.logs.size.mb=4096, pacemaker.childopts=-Xmx1024m, ui.users=null, transactional.zookeeper.servers=null, supervisor.worker.timeout.secs=30, storm.zookeeper.auth.password=null, storm.blobstore.acl.validation.enabled=false, client.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.thrift.socket.timeout.ms=600000, supervisor.childopts=-Xmx256m, topology.worker.max.heap.size.mb=768.0, ui.http.x-frame-options=DENY, backpressure.disruptor.high.watermark=0.9, ui.filter=null, ui.header.buffer.bytes=4096, topology.min.replication.count=1, topology.disruptor.wait.timeout.millis=1000, storm.nimbus.retry.intervalceiling.millis=60000, topology.trident.batch.emit.interval.millis=50, storm.auth.simple-acl.users=[], drpc.invocations.threads=64, java.library.path=/usr/local/lib:/opt/local/lib:/usr/lib, ui.port=8080, storm.exhibitor.poll.uripath=/exhibitor/v1/cluster/list, storm.messaging.netty.transfer.batch.size=262144, logviewer.appender.name=A1, nimbus.thrift.max_buffer_size=1048576, storm.auth.simple-acl.users.commands=[], drpc.request.timeout.secs=600}

4750 [main] WARN  o.a.s.d.s.Slot - SLOT ip-172-31-22-101.us-west-2.compute.internal:1027 Starting in state EMPTY - assignment null

4750 [main] WARN  o.a.s.d.s.Slot - SLOT ip-172-31-22-101.us-west-2.compute.internal:1028 Starting in state EMPTY - assignment null

4750 [main] WARN  o.a.s.d.s.Slot - SLOT ip-172-31-22-101.us-west-2.compute.internal:1029 Starting in state EMPTY - assignment null

4750 [main] INFO  o.a.s.l.AsyncLocalizer - Cleaning up unused topologies in /tmp/a84faee2-1f19-46b6-affa-3c9d4bb2d61a/supervisor/stormdist

4751 [main] INFO  o.a.s.d.s.Supervisor - Starting supervisor with id d9f039bd-9a40-4a5b-ae52-b2638fbedaf8 at host ip-172-31-22-101.us-west-2.compute.internal.

4759 [main] WARN  o.a.s.u.Utils - STORM-VERSION new 1.2.3 old 1.2.3

4850 [main] INFO  o.a.s.d.nimbus - Received topology submission for crawler (storm-1.2.3 JDK-11.0.4) with conf {parser.emitOutlinks.max.per.page=-1, es.metrics.index.name=metrics, topology.max.task.parallelism=null, status.updater.cache.spec=maximumSize=10000,expireAfterAccess=1h, fetcher.queue.mode=byHost, http.agent.email=, status.updater.use.cache=true, partition.url.mode=byHost, status.updater.unit.round.date=SECOND, max.fetch.errors=3, selenium.implicitlyWait=0, robots.error.cache.spec=maximumSize=10000,expireAfterWrite=1h, detect.charset.maxlength=10000, es.status.concurrentRequests=1, es.status.routing.fieldname=key, selenium.pageLoadTimeout=-1, indexer.url.fieldname=url, es.indexer.addresses=localhost, fetchInterval.fetch.error=120, spout.reset.fetchdate.after=120, fetcher.max.urls.in.queues=-1, fetchInterval.default=1440, es.status.recentDate.min.gap=-1, http.agent.version=0.1, urlbuffer.class=com.digitalpebble.stormcrawler.persistence.SimpleURLBuffer, es.status.max.start.offset=500, http.agent.description=, es.status.addresses=http://localhost:9200, topology.submitter.principal=, topology.acker.executors=null, topology.eventlogger.executors=0, http.content.partial.as.trimmed=false, topology.worker.childopts=-Xmx2g -Djava.net.preferIPv4Stack=true, file.protocol.implementation=com.digitalpebble.stormcrawler.protocol.file.FileProtocol, http.timeout=10000, es.status.bucket.sort.field=["nextFetchDate" "url"], fetcher.timeout.queue=-1, http.agent.name=Google-Read-Aloud, topology.workers=1, jsoup.treat.non.html.as.error=true, topology.message.timeout.secs=300, topology.debug=false, sitemap.discovery=false, spout.min.delay.queries=2000, sitemap.filter.hours.since.modified=-1, textextractor.include.pattern=["DIV[id=\"maincontent\"]" "DIV[itemprop=\"articleBody\"]" "ARTICLE"], topology.max.spout.pending=100, fetcher.threads.per.queue=1, protocols=http,https,file, fetcher.server.delay.force=false, storm.zookeeper.superACL=null, es.indexer.settings={"cluster.name" "elasticsearch"}, indexer.text.fieldname=content, http.accept.language=en-us,en-gb,en;q=0.7,*;q=0.3, es.status.sample=false, http.robots.403.allow=true, http.store.headers=false, fetcher.server.delay=1.0, http.agent.url=https://support.google.com/webmasters/answer/1061943, fetcher.metrics.time.bucket.secs=10, es.status.index.name=status, spout.ttl.purgatory=30, detect.mimetype=true, fetcher.max.crawl.delay.force=false, sitemap.schedule.delay=-1, metadata.track.depth=true, es.status.settings={"cluster.name" "elasticsearch"}, es.status.global.sort.field=nextFetchDate, es.indexer.create=false, topology.users=clojure.lang.LazySeq@1, track.anchors=true, topology.submitter.user=ubuntu, metadata.track.path=true, http.content.limit=65536, http.accept=text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8, topology.metrics.consumer.register=[{"class" "com.digitalpebble.stormcrawler.elasticsearch.metrics.MetricsConsumer", "parallelism.hint" 1}], fetcher.max.crawl.delay=30, es.status.bulkActions=500, es.indexer.flushInterval=2s, es.status.flushInterval=5s, selenium.delegated.protocol=com.digitalpebble.stormcrawler.protocol.httpclient.HttpProtocol, es.indexer.bulkActions=100, topology.kryo.register={"com.digitalpebble.stormcrawler.Metadata" nil}, urlfilters.config.file=urlfilters.json, es.status.query.timeout=-1, metadata.persist=["_redirTo" "error.cause" "error.source" "isSitemap" "isFeed"], parser.emitOutlinks=true, textextractor.exclude.tags=["STYLE" "SCRIPT"], indexer.text.maxlength=-1, fetcher.threads.number=50, es.metrics.settings={"cluster.name" "elasticsearch"}, selenium.instances.num=1, selenium.capabilities={"loadImages" false, "takesScreenshot" false, "javascriptEnabled" true}, scheduler.class=com.digitalpebble.stormcrawler.persistence.DefaultScheduler, robots.cache.spec=maximumSize=10000,expireAfterWrite=6h, es.metrics.addresses=http://localhost:9200, es.indexer.concurrentRequests=1, robots.noFollow.strict=true, indexer.md.mapping=["parse.title=title" "parse.keywords=keywords" "parse.description=description" "domain=domain"], es.status.recentDate.increase=-1, https.protocol.implementation=com.digitalpebble.stormcrawler.protocol.httpclient.HttpProtocol, es.status.max.urls.per.bucket=2, es.indexer.index.name=content, fetcher.max.queue.size=-1, selenium.setScriptTimeout=0, indexer.canonical.name=canonical, es.status.bucket.field=key, es.status.routing=true, topology.kryo.decorators=clojure.lang.LazySeq@1, es.status.max.buckets=50, storm.id=crawler-1-1574269039, topology.name=crawler, fetcher.max.throttle.sleep=-1, fetcher.server.min.delay=0.0, parsefilters.config.file=parsefilters.json, http.protocol.implementation=com.digitalpebble.stormcrawler.protocol.httpclient.HttpProtocol, fetchInterval.error=-1}

4856 [main] INFO  o.a.s.d.nimbus - uploadedJar 

4866 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f20000 type:create cxid:0x9 zxid:0x25 txntype:-1 reqpath:n/a Error Path:/storm/blobstoremaxkeysequencenumber Error:KeeperErrorCode = NoNode for /storm/blobstoremaxkeysequencenumber

4874 [main] INFO  o.a.s.cluster - setup-path/blobstore/crawler-1-1574269039-stormconf.ser/ip-172-31-22-101.us-west-2.compute.internal:6627-1

4919 [main] INFO  o.a.s.cluster - setup-path/blobstore/crawler-1-1574269039-stormcode.ser/ip-172-31-22-101.us-west-2.compute.internal:6627-1

4947 [main] INFO  o.a.s.d.nimbus - desired replication count 1 achieved, current-replication-count for conf key = 1, current-replication-count for code key = 1, current-replication-count for jar key = 1

5004 [main] INFO  o.a.s.d.nimbus - Activating crawler: crawler-1-1574269039

5502 [timer] INFO  o.a.s.s.EvenScheduler - Available slots: (["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1025] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1026] ["d9f039bd-9a40-4a5b-ae52-b2638fbedaf8" 1027] ["d9f039bd-9a40-4a5b-ae52-b2638fbedaf8" 1028] ["d9f039bd-9a40-4a5b-ae52-b2638fbedaf8" 1029])

5535 [timer] INFO  o.a.s.d.nimbus - Setting new assignment for topology id crawler-1-1574269039: #org.apache.storm.daemon.common.Assignment{:master-code-dir "/tmp/55b9f681-e0ff-4929-81ff-d3354e5f786e", :node->host {"ab823f98-11b3-4021-b8e7-f9da7ebc12f3" "ip-172-31-22-101.us-west-2.compute.internal"}, :executor->node+port {[8 8] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [12 12] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [2 2] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [7 7] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [3 3] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [1 1] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [18 18] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [6 6] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [9 9] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [11 11] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [16 16] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [13 13] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [19 19] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [5 5] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [10 10] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [14 14] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [4 4] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [15 15] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024], [17 17] ["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024]}, :executor->start-time-secs {[8 8] 1574269040, [12 12] 1574269040, [2 2] 1574269040, [7 7] 1574269040, [3 3] 1574269040, [1 1] 1574269040, [18 18] 1574269040, [6 6] 1574269040, [9 9] 1574269040, [11 11] 1574269040, [16 16] 1574269040, [13 13] 1574269040, [19 19] 1574269040, [5 5] 1574269040, [10 10] 1574269040, [14 14] 1574269040, [4 4] 1574269040, [15 15] 1574269040, [17 17] 1574269040}, :worker->resources {["ab823f98-11b3-4021-b8e7-f9da7ebc12f3" 1024] [0.0 0.0 0.0]}, :owner "ubuntu"}

5712 [SLOT_1024] INFO  o.a.s.d.s.Slot - STATE EMPTY msInState: 1008 -> WAITING_FOR_BASIC_LOCALIZATION msInState: 0

5712 [Async Localizer] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

5725 [Async Localizer] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@4a59bca0

5726 [Async Localizer-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

5727 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55336

5727 [Async Localizer-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

5728 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55336

5728 [Async Localizer] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

5728 [Async Localizer] INFO  o.a.s.b.FileBlobStoreImpl - Creating new blob store based in /tmp/55b9f681-e0ff-4929-81ff-d3354e5f786e/blobs

5729 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f2000d with negotiated timeout 20000 for client /127.0.0.1:55336

5730 [Async Localizer-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f2000d, negotiated timeout = 20000

5730 [Async Localizer-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

5736 [Curator-Framework-0] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting

5740 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x100040091f2000d

5741 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:55336 which had sessionid 0x100040091f2000d

5741 [Async Localizer] INFO  o.a.s.s.o.a.z.ZooKeeper - Session: 0x100040091f2000d closed

5741 [Async Localizer-EventThread] INFO  o.a.s.s.o.a.z.ClientCnxn - EventThread shut down for session: 0x100040091f2000d

5790 [Async Localizer] INFO  o.a.s.l.AsyncLocalizer - Extracting resources from jar at target/dev1-0.1.jar to /tmp/ddf53adf-8313-4830-8786-4bd66e78b4fe/supervisor/tmp/9bbef7a4-772d-4a8e-8c37-b14acb2718ee/

5804 [SLOT_1024] INFO  o.a.s.d.s.Slot - STATE WAITING_FOR_BASIC_LOCALIZATION msInState: 92 -> WAITING_FOR_BLOB_LOCALIZATION msInState: 0

5811 [SLOT_1024] INFO  o.a.s.d.s.Container - Setting up ab823f98-11b3-4021-b8e7-f9da7ebc12f3:f5d45862-968e-43f7-97e8-669e3dae7d11

5816 [SLOT_1024] INFO  o.a.s.d.s.Container - GET worker-user for f5d45862-968e-43f7-97e8-669e3dae7d11

5826 [SLOT_1024] INFO  o.a.s.d.s.Container - SET worker-user f5d45862-968e-43f7-97e8-669e3dae7d11 ubuntu

5832 [SLOT_1024] INFO  o.a.s.d.worker - Launching worker for crawler-1-1574269039 on ab823f98-11b3-4021-b8e7-f9da7ebc12f3:1024 with id f5d45862-968e-43f7-97e8-669e3dae7d11 and conf {topology.builtin.metrics.bucket.size.secs=60, nimbus.childopts=-Xmx1024m, ui.filter.params=null, storm.cluster.mode=local, storm.messaging.netty.client_worker_threads=1, logviewer.max.per.worker.logs.size.mb=2048, supervisor.run.worker.as.user=false, topology.max.task.parallelism=null, topology.priority=29, zmq.threads=1, storm.group.mapping.service=org.apache.storm.security.auth.ShellBasedGroupsMapping, transactional.zookeeper.root=/transactional, topology.sleep.spout.wait.strategy.time.ms=1, ui.pagination=20, scheduler.display.resource=false, topology.max.replication.wait.time.sec=60, drpc.invocations.port=3773, supervisor.localizer.cache.target.size.mb=10240, topology.multilang.serializer=org.apache.storm.multilang.JsonSerializer, storm.messaging.netty.server_worker_threads=1, nimbus.blobstore.class=org.apache.storm.blobstore.LocalFsBlobStore, resource.aware.scheduler.eviction.strategy=org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy, topology.max.error.report.per.interval=5, storm.thrift.transport=org.apache.storm.security.auth.SimpleTransportPlugin, zmq.hwm=0, storm.group.mapping.service.params=null, worker.profiler.enabled=false, storm.principal.tolocal=org.apache.storm.security.auth.DefaultPrincipalToLocal, supervisor.worker.shutdown.sleep.secs=3, pacemaker.host=localhost, storm.zookeeper.retry.times=5, ui.actions.enabled=true, zmq.linger.millis=0, supervisor.enable=true, topology.stats.sample.rate=0.05, storm.nimbus.zookeeper.acls.fixup=true, storm.messaging.netty.min_wait_ms=100, worker.log.level.reset.poll.secs=30, storm.zookeeper.port=2000, supervisor.heartbeat.frequency.secs=5, topology.enable.message.timeouts=true, supervisor.cpu.capacity=400.0, drpc.worker.threads=64, supervisor.blobstore.download.thread.count=5, task.backpressure.poll.secs=30, drpc.queue.size=128, topology.backpressure.enable=false, supervisor.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.blobstore.inputstream.buffer.size.bytes=65536, topology.shellbolt.max.pending=100, drpc.https.keystore.password=*****, nimbus.code.sync.freq.secs=120, logviewer.port=8000, topology.scheduler.strategy=org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy, topology.executor.send.buffer.size=1024, resource.aware.scheduler.priority.strategy=org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy, pacemaker.auth.method=NONE, storm.daemon.metrics.reporter.plugins=["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], topology.worker.logwriter.childopts=-Xmx64m, topology.spout.wait.strategy=org.apache.storm.spout.SleepSpoutWaitStrategy, ui.host=0.0.0.0, storm.nimbus.retry.interval.millis=2000, nimbus.inbox.jar.expiration.secs=3600, dev.zookeeper.path=/tmp/dev-storm-zookeeper, topology.acker.executors=null, topology.fall.back.on.java.serialization=true, topology.eventlogger.executors=0, supervisor.localizer.cleanup.interval.ms=600000, storm.zookeeper.servers=["localhost"], nimbus.thrift.threads=64, logviewer.cleanup.age.mins=10080, topology.worker.childopts=null, topology.classpath=null, supervisor.monitor.frequency.secs=3, nimbus.credential.renewers.freq.secs=600, backpressure.znode.timeout.secs=30, topology.skip.missing.kryo.registrations=true, drpc.authorizer.acl.filename=drpc-auth-acl.yaml, pacemaker.kerberos.users=[], storm.group.mapping.service.cache.duration.secs=120, blobstore.dir=/tmp/55b9f681-e0ff-4929-81ff-d3354e5f786e, topology.testing.always.try.serialize=false, nimbus.monitor.freq.secs=10, storm.health.check.timeout.ms=5000, supervisor.supervisors=[], topology.tasks=null, topology.bolts.outgoing.overflow.buffer.enable=false, storm.messaging.netty.socket.backlog=500, topology.workers=1, pacemaker.base.threads=10, storm.local.dir=/tmp/ddf53adf-8313-4830-8786-4bd66e78b4fe, backpressure.znode.update.freq.secs=15, worker.childopts=-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump, storm.auth.simple-white-list.users=[], topology.disruptor.batch.timeout.millis=1, topology.message.timeout.secs=30, topology.state.synchronization.timeout.secs=60, topology.tuple.serializer=org.apache.storm.serialization.types.ListDelegateSerializer, supervisor.supervisors.commands=[], nimbus.blobstore.expiration.secs=600, logviewer.childopts=-Xmx128m, storm.nimbus.zookeeper.acls.check=true, topology.environment=null, topology.debug=false, topology.disruptor.batch.size=100, storm.disable.symlinks=false, storm.messaging.netty.max_retries=300, ui.childopts=-Xmx768m, storm.network.topography.plugin=org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping, storm.zookeeper.session.timeout=20000, drpc.childopts=-Xmx768m, drpc.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.connection.timeout=15000, storm.zookeeper.auth.user=null, storm.meta.serialization.delegate=org.apache.storm.serialization.GzipThriftSerializationDelegate, topology.max.spout.pending=null, storm.codedistributor.class=org.apache.storm.codedistributor.LocalFileSystemCodeDistributor, nimbus.supervisor.timeout.secs=60, nimbus.task.timeout.secs=30, drpc.port=3772, pacemaker.max.threads=50, storm.zookeeper.retry.intervalceiling.millis=30000, nimbus.thrift.port=6627, storm.auth.simple-acl.admins=[], topology.component.cpu.pcore.percent=10.0, supervisor.memory.capacity.mb=3072.0, storm.nimbus.retry.times=5, supervisor.worker.start.timeout.secs=120, storm.zookeeper.retry.interval=1000, logs.users=null, storm.cluster.metrics.consumer.publish.interval.secs=60, worker.profiler.command=flight.bash, transactional.zookeeper.port=null, drpc.max_buffer_size=1048576, pacemaker.thread.timeout=10, task.credentials.poll.secs=30, blobstore.superuser=ubuntu, drpc.https.keystore.type=JKS, topology.worker.receiver.thread.count=1, topology.state.checkpoint.interval.ms=1000, supervisor.slots.ports=clojure.lang.LazySeq@ff880, topology.transfer.buffer.size=1024, storm.health.check.dir=healthchecks, topology.worker.shared.thread.pool.size=4, drpc.authorizer.acl.strict=false, nimbus.file.copy.expiration.secs=600, worker.profiler.childopts=-XX:+UnlockCommercialFeatures -XX:+FlightRecorder, topology.executor.receive.buffer.size=1024, backpressure.disruptor.low.watermark=0.4, nimbus.task.launch.secs=120, storm.local.mode.zmq=false, storm.messaging.netty.buffer_size=5242880, storm.cluster.state.store=org.apache.storm.cluster_state.zookeeper_state_factory, worker.heartbeat.frequency.secs=1, storm.log4j2.conf.dir=log4j2, ui.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.root=/storm, topology.tick.tuple.freq.secs=null, drpc.https.port=-1, storm.workers.artifacts.dir=workers-artifacts, supervisor.blobstore.download.max_retries=3, task.refresh.poll.secs=10, storm.exhibitor.port=8080, task.heartbeat.frequency.secs=3, pacemaker.port=6699, storm.messaging.netty.max_wait_ms=1000, topology.component.resources.offheap.memory.mb=0.0, drpc.http.port=3774, topology.error.throttle.interval.secs=10, storm.messaging.transport=org.apache.storm.messaging.netty.Context, topology.disable.loadaware.messaging=false, storm.messaging.netty.authentication=false, topology.component.resources.onheap.memory.mb=128.0, topology.kryo.factory=org.apache.storm.serialization.DefaultKryoFactory, worker.gc.childopts=, nimbus.topology.validator=org.apache.storm.nimbus.DefaultTopologyValidator, nimbus.seeds=["localhost"], nimbus.queue.size=100000, nimbus.cleanup.inbox.freq.secs=600, storm.blobstore.replication.factor=3, worker.heap.memory.mb=768, logviewer.max.sum.worker.logs.size.mb=4096, pacemaker.childopts=-Xmx1024m, ui.users=null, transactional.zookeeper.servers=null, supervisor.worker.timeout.secs=30, storm.zookeeper.auth.password=null, storm.blobstore.acl.validation.enabled=false, client.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.thrift.socket.timeout.ms=600000, supervisor.childopts=-Xmx256m, topology.worker.max.heap.size.mb=768.0, ui.http.x-frame-options=DENY, backpressure.disruptor.high.watermark=0.9, ui.filter=null, ui.header.buffer.bytes=4096, topology.min.replication.count=1, topology.disruptor.wait.timeout.millis=1000, storm.nimbus.retry.intervalceiling.millis=60000, topology.trident.batch.emit.interval.millis=50, storm.auth.simple-acl.users=[], drpc.invocations.threads=64, java.library.path=/usr/local/lib:/opt/local/lib:/usr/lib, ui.port=8080, storm.exhibitor.poll.uripath=/exhibitor/v1/cluster/list, storm.messaging.netty.transfer.batch.size=262144, logviewer.appender.name=A1, nimbus.thrift.max_buffer_size=1048576, storm.auth.simple-acl.users.commands=[], drpc.request.timeout.secs=600}

5834 [SLOT_1024] INFO  o.a.s.m.StormMetricRegistry - Starting metrics reporters...

5839 [SLOT_1024] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

5840 [SLOT_1024] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@79eb66d

5848 [SLOT_1024] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

5852 [SLOT_1024-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

5852 [SLOT_1024-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

5852 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55338

5853 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55338

5854 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f2000e with negotiated timeout 20000 for client /127.0.0.1:55338

5855 [SLOT_1024-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f2000e, negotiated timeout = 20000

5855 [SLOT_1024-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

5856 [SLOT_1024-EventThread] INFO  o.a.s.zookeeper - Zookeeper state update: :connected:none

5859 [Curator-Framework-0] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting

5860 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x100040091f2000e

5861 [SLOT_1024] INFO  o.a.s.s.o.a.z.ZooKeeper - Session: 0x100040091f2000e closed

5861 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:55338 which had sessionid 0x100040091f2000e

5862 [SLOT_1024] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting

5862 [SLOT_1024-EventThread] INFO  o.a.s.s.o.a.z.ClientCnxn - EventThread shut down for session: 0x100040091f2000e

5863 [SLOT_1024] INFO  o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@2ded3962

5869 [SLOT_1024] INFO  o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Default schema

5873 [SLOT_1024-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)

5873 [SLOT_1024-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Socket connection established to localhost/127.0.0.1:2000, initiating session

5873 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:55340

5873 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:55340

5875 [SyncThread:0] INFO  o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x100040091f2000f with negotiated timeout 20000 for client /127.0.0.1:55340

5876 [SLOT_1024-SendThread(localhost:2000)] INFO  o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x100040091f2000f, negotiated timeout = 20000

5876 [SLOT_1024-EventThread] INFO  o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED

5882 [SLOT_1024] INFO  o.a.s.s.a.AuthUtils - Got AutoCreds []

5885 [SLOT_1024] INFO  o.a.s.d.worker - Reading Assignments.

6022 [SLOT_1024] INFO  o.a.s.d.worker - Registering IConnectionCallbacks for ab823f98-11b3-4021-b8e7-f9da7ebc12f3:1024

6074 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor spout:[8 8]

6091 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks spout:[8 8]

WARNING: An illegal reflective access operation has occurred

WARNING: Illegal reflective access by com.esotericsoftware.kryo.util.UnsafeUtil (file:/opt/apache-storm-1.2.3/lib/kryo-3.0.3.jar) to constructor java.nio.DirectByteBuffer(long,int,java.lang.Object)

WARNING: Please consider reporting this to the maintainers of com.esotericsoftware.kryo.util.UnsafeUtil

WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations

WARNING: All illegal access operations will be denied in a future release

6224 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor spout:[8 8]

6235 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor spout:[12 12]

6236 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks spout:[12 12]

6250 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor spout:[12 12]

6257 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor __metricscom.digitalpebble.stormcrawler.elasticsearch.metrics.MetricsConsumer:[2 2]

6264 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks __metricscom.digitalpebble.stormcrawler.elasticsearch.metrics.MetricsConsumer:[2 2]

6269 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor __metricscom.digitalpebble.stormcrawler.elasticsearch.metrics.MetricsConsumer:[2 2]

6275 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor sitemap:[7 7]

6276 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks sitemap:[7 7]

6279 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor sitemap:[7 7]

6285 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor fetcher:[3 3]

6285 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks fetcher:[3 3]

6292 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor fetcher:[3 3]

6299 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor __acker:[1 1]

6300 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks __acker:[1 1]

6302 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor __acker:[1 1]

6308 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor status:[18 18]

6309 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks status:[18 18]

6312 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor status:[18 18]

6319 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor partitioner:[6 6]

6320 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks partitioner:[6 6]

6331 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor partitioner:[6 6]

6337 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor spout:[9 9]

6338 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks spout:[9 9]

6347 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor spout:[9 9]

6354 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor spout:[11 11]

6356 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks spout:[11 11]

6366 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor spout:[11 11]

6373 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor __system:[-1 -1]

6373 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks __system:[-1 -1]

6379 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor __system:[-1 -1]

6389 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor spout:[16 16]

6391 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks spout:[16 16]

6399 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor spout:[16 16]

6407 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor spout:[13 13]

6410 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks spout:[13 13]

6413 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor spout:[13 13]

6423 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor status_metrics:[19 19]

6423 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks status_metrics:[19 19]

6430 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor status_metrics:[19 19]

6438 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor parse:[5 5]

6520 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks parse:[5 5]

6526 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor parse:[5 5]

6536 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor spout:[10 10]

6537 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks spout:[10 10]

6546 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor spout:[10 10]

6553 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor spout:[14 14]

6554 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks spout:[14 14]

6562 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor spout:[14 14]

6568 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor index:[4 4]

6600 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks index:[4 4]

6606 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor index:[4 4]

6613 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor spout:[15 15]

6614 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks spout:[15 15]

6618 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor spout:[15 15]

6625 [SLOT_1024] INFO  o.a.s.d.executor - Loading executor spout:[17 17]

6626 [SLOT_1024] INFO  o.a.s.d.executor - Loaded executor tasks spout:[17 17]

6630 [SLOT_1024] INFO  o.a.s.d.executor - Finished loading executor spout:[17 17]

6640 [SLOT_1024] INFO  o.a.s.d.worker - Started with log levels: {"" #object[org.apache.logging.log4j.Level 0x2bcfe0cd "INFO"], "org.apache.zookeeper" #object[org.apache.logging.log4j.Level 0x29b0533e "WARN"]}

6648 [SLOT_1024] INFO  o.a.s.d.worker - Worker has topology config {topology.builtin.metrics.bucket.size.secs=60, nimbus.childopts=-Xmx1024m, ui.filter.params=null, storm.cluster.mode=local, parser.emitOutlinks.max.per.page=-1, storm.messaging.netty.client_worker_threads=1, logviewer.max.per.worker.logs.size.mb=2048, supervisor.run.worker.as.user=false, es.metrics.index.name=metrics, topology.max.task.parallelism=null, status.updater.cache.spec=maximumSize=10000,expireAfterAccess=1h, topology.priority=29, zmq.threads=1, storm.group.mapping.service=org.apache.storm.security.auth.ShellBasedGroupsMapping, transactional.zookeeper.root=/transactional, fetcher.queue.mode=byHost, http.agent.email=, topology.sleep.spout.wait.strategy.time.ms=1, ui.pagination=20, status.updater.use.cache=true, scheduler.display.resource=false, topology.max.replication.wait.time.sec=60, drpc.invocations.port=3773, supervisor.localizer.cache.target.size.mb=10240, topology.multilang.serializer=org.apache.storm.multilang.JsonSerializer, storm.messaging.netty.server_worker_threads=1, nimbus.blobstore.class=org.apache.storm.blobstore.LocalFsBlobStore, resource.aware.scheduler.eviction.strategy=org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy, partition.url.mode=byHost, status.updater.unit.round.date=SECOND, max.fetch.errors=3, topology.max.error.report.per.interval=5, storm.thrift.transport=org.apache.storm.security.auth.SimpleTransportPlugin, zmq.hwm=0, storm.group.mapping.service.params=null, worker.profiler.enabled=false, selenium.implicitlyWait=0, robots.error.cache.spec=maximumSize=10000,expireAfterWrite=1h, detect.charset.maxlength=10000, storm.principal.tolocal=org.apache.storm.security.auth.DefaultPrincipalToLocal, supervisor.worker.shutdown.sleep.secs=3, pacemaker.host=localhost, storm.zookeeper.retry.times=5, ui.actions.enabled=true, es.status.concurrentRequests=1, es.status.routing.fieldname=key, zmq.linger.millis=0, supervisor.enable=true, topology.stats.sample.rate=0.05, selenium.pageLoadTimeout=-1, indexer.url.fieldname=url, es.indexer.addresses=localhost, storm.nimbus.zookeeper.acls.fixup=true, storm.messaging.netty.min_wait_ms=100, fetchInterval.fetch.error=120, worker.log.level.reset.poll.secs=30, storm.zookeeper.port=2000, supervisor.heartbeat.frequency.secs=5, spout.reset.fetchdate.after=120, topology.enable.message.timeouts=true, supervisor.cpu.capacity=400.0, fetcher.max.urls.in.queues=-1, fetchInterval.default=1440, drpc.worker.threads=64, supervisor.blobstore.download.thread.count=5, task.backpressure.poll.secs=30, drpc.queue.size=128, es.status.recentDate.min.gap=-1, http.agent.version=0.1, urlbuffer.class=com.digitalpebble.stormcrawler.persistence.SimpleURLBuffer, es.status.max.start.offset=500, topology.backpressure.enable=false, supervisor.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.blobstore.inputstream.buffer.size.bytes=65536, topology.shellbolt.max.pending=100, http.agent.description=, drpc.https.keystore.password=*****, nimbus.code.sync.freq.secs=120, logviewer.port=8000, topology.scheduler.strategy=org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy, topology.executor.send.buffer.size=1024, resource.aware.scheduler.priority.strategy=org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy, pacemaker.auth.method=NONE, storm.daemon.metrics.reporter.plugins=["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], es.status.addresses=http://localhost:9200, topology.worker.logwriter.childopts=-Xmx64m, topology.spout.wait.strategy=org.apache.storm.spout.SleepSpoutWaitStrategy, ui.host=0.0.0.0, topology.submitter.principal=, storm.nimbus.retry.interval.millis=2000, nimbus.inbox.jar.expiration.secs=3600, dev.zookeeper.path=/tmp/dev-storm-zookeeper, topology.acker.executors=null, topology.fall.back.on.java.serialization=true, topology.eventlogger.executors=0, http.content.partial.as.trimmed=false, supervisor.localizer.cleanup.interval.ms=600000, storm.zookeeper.servers=["localhost"], nimbus.thrift.threads=64, logviewer.cleanup.age.mins=10080, topology.worker.childopts=-Xmx2g -Djava.net.preferIPv4Stack=true, topology.classpath=null, supervisor.monitor.frequency.secs=3, file.protocol.implementation=com.digitalpebble.stormcrawler.protocol.file.FileProtocol, nimbus.credential.renewers.freq.secs=600, backpressure.znode.timeout.secs=30, http.timeout=10000, topology.skip.missing.kryo.registrations=true, drpc.authorizer.acl.filename=drpc-auth-acl.yaml, es.status.bucket.sort.field=["nextFetchDate" "url"], pacemaker.kerberos.users=[], fetcher.timeout.queue=-1, http.agent.name=Google-Read-Aloud, storm.group.mapping.service.cache.duration.secs=120, blobstore.dir=/tmp/55b9f681-e0ff-4929-81ff-d3354e5f786e, topology.testing.always.try.serialize=false, nimbus.monitor.freq.secs=10, storm.health.check.timeout.ms=5000, supervisor.supervisors=[], topology.tasks=null, topology.bolts.outgoing.overflow.buffer.enable=false, storm.messaging.netty.socket.backlog=500, topology.workers=1, pacemaker.base.threads=10, storm.local.dir=/tmp/55b9f681-e0ff-4929-81ff-d3354e5f786e, backpressure.znode.update.freq.secs=15, worker.childopts=-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump, storm.auth.simple-white-list.users=[], topology.disruptor.batch.timeout.millis=1, jsoup.treat.non.html.as.error=true, topology.message.timeout.secs=300, topology.state.synchronization.timeout.secs=60, topology.tuple.serializer=org.apache.storm.serialization.types.ListDelegateSerializer, supervisor.supervisors.commands=[], nimbus.blobstore.expiration.secs=600, logviewer.childopts=-Xmx128m, storm.nimbus.zookeeper.acls.check=true, topology.environment=null, topology.debug=false, sitemap.discovery=false, spout.min.delay.queries=2000, topology.disruptor.batch.size=100, storm.disable.symlinks=false, storm.messaging.netty.max_retries=300, ui.childopts=-Xmx768m, storm.network.topography.plugin=org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping, storm.zookeeper.session.timeout=20000, drpc.childopts=-Xmx768m, drpc.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.connection.timeout=15000, storm.zookeeper.auth.user=null, sitemap.filter.hours.since.modified=-1, textextractor.include.pattern=["DIV[id=\"maincontent\"]" "DIV[itemprop=\"articleBody\"]" "ARTICLE"], storm.meta.serialization.delegate=org.apache.storm.serialization.GzipThriftSerializationDelegate, topology.max.spout.pending=100, storm.codedistributor.class=org.apache.storm.codedistributor.LocalFileSystemCodeDistributor, fetcher.threads.per.queue=1, nimbus.supervisor.timeout.secs=60, protocols=http,https,file, fetcher.server.delay.force=false, nimbus.task.timeout.secs=30, storm.zookeeper.superACL=null, drpc.port=3772, pacemaker.max.threads=50, storm.zookeeper.retry.intervalceiling.millis=30000, es.indexer.settings={"cluster.name" "elasticsearch"}, nimbus.thrift.port=6627, storm.auth.simple-acl.admins=[], topology.component.cpu.pcore.percent=10.0, supervisor.memory.capacity.mb=3072.0, indexer.text.fieldname=content, http.accept.language=en-us,en-gb,en;q=0.7,*;q=0.3, storm.nimbus.retry.times=5, supervisor.worker.start.timeout.secs=120, es.status.sample=false, http.robots.403.allow=true, http.store.headers=false, storm.zookeeper.retry.interval=1000, fetcher.server.delay=1.0, http.agent.url=https://support.google.com/webmasters/answer/1061943, fetcher.metrics.time.bucket.secs=10, es.status.index.name=status, logs.users=null, storm.cluster.metrics.consumer.publish.interval.secs=60, worker.profiler.command=flight.bash, transactional.zookeeper.port=null, drpc.max_buffer_size=1048576, pacemaker.thread.timeout=10, spout.ttl.purgatory=30, task.credentials.poll.secs=30, blobstore.superuser=ubuntu, drpc.https.keystore.type=JKS, topology.worker.receiver.thread.count=1, topology.state.checkpoint.interval.ms=1000, supervisor.slots.ports=[6700 6701 6702 6703], detect.mimetype=true, fetcher.max.crawl.delay.force=false, topology.transfer.buffer.size=1024, sitemap.schedule.delay=-1, storm.health.check.dir=healthchecks, topology.worker.shared.thread.pool.size=4, drpc.authorizer.acl.strict=false, nimbus.file.copy.expiration.secs=600, metadata.track.depth=true, es.status.settings={"cluster.name" "elasticsearch"}, es.status.global.sort.field=nextFetchDate, es.indexer.create=false, worker.profiler.childopts=-XX:+UnlockCommercialFeatures -XX:+FlightRecorder, topology.executor.receive.buffer.size=1024, backpressure.disruptor.low.watermark=0.4, topology.users=[], nimbus.task.launch.secs=120, storm.local.mode.zmq=false, track.anchors=true, storm.messaging.netty.buffer_size=5242880, storm.cluster.state.store=org.apache.storm.cluster_state.zookeeper_state_factory, worker.heartbeat.frequency.secs=1, storm.log4j2.conf.dir=log4j2, ui.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.root=/storm, topology.submitter.user=ubuntu, topology.tick.tuple.freq.secs=null, metadata.track.path=true, drpc.https.port=-1, http.content.limit=65536, storm.workers.artifacts.dir=workers-artifacts, http.accept=text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8, supervisor.blobstore.download.max_retries=3, task.refresh.poll.secs=10, topology.metrics.consumer.register=[{"class" "com.digitalpebble.stormcrawler.elasticsearch.metrics.MetricsConsumer", "parallelism.hint" 1}], storm.exhibitor.port=8080, task.heartbeat.frequency.secs=3, fetcher.max.crawl.delay=30, pacemaker.port=6699, storm.messaging.netty.max_wait_ms=1000, es.status.bulkActions=500, topology.component.resources.offheap.memory.mb=0.0, drpc.http.port=3774, es.indexer.flushInterval=2s, topology.error.throttle.interval.secs=10, storm.messaging.transport=org.apache.storm.messaging.netty.Context, es.status.flushInterval=5s, topology.disable.loadaware.messaging=false, selenium.delegated.protocol=com.digitalpebble.stormcrawler.protocol.httpclient.HttpProtocol, storm.messaging.netty.authentication=false, topology.component.resources.onheap.memory.mb=128.0, es.indexer.bulkActions=100, topology.kryo.factory=org.apache.storm.serialization.DefaultKryoFactory, topology.kryo.register={"com.digitalpebble.stormcrawler.Metadata" nil}, urlfilters.config.file=urlfilters.json, worker.gc.childopts=, nimbus.topology.validator=org.apache.storm.nimbus.DefaultTopologyValidator, es.status.query.timeout=-1, nimbus.seeds=["localhost"], metadata.persist=["_redirTo" "error.cause" "error.source" "isSitemap" "isFeed"], nimbus.queue.size=100000, nimbus.cleanup.inbox.freq.secs=600, storm.blobstore.replication.factor=3, worker.heap.memory.mb=768, logviewer.max.sum.worker.logs.size.mb=4096, pacemaker.childopts=-Xmx1024m, ui.users=null, parser.emitOutlinks=true, transactional.zookeeper.servers=null, supervisor.worker.timeout.secs=30, storm.zookeeper.auth.password=null, textextractor.exclude.tags=["STYLE" "SCRIPT"], storm.blobstore.acl.validation.enabled=false, client.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, indexer.text.maxlength=-1, storm.thrift.socket.timeout.ms=600000, supervisor.childopts=-Xmx256m, fetcher.threads.number=50, topology.worker.max.heap.size.mb=768.0, ui.http.x-frame-options=DENY, backpressure.disruptor.high.watermark=0.9, ui.filter=null, es.metrics.settings={"cluster.name" "elasticsearch"}, selenium.instances.num=1, ui.header.buffer.bytes=4096, topology.min.replication.count=1, selenium.capabilities={"loadImages" false, "takesScreenshot" false, "javascriptEnabled" true}, scheduler.class=com.digitalpebble.stormcrawler.persistence.DefaultScheduler, robots.cache.spec=maximumSize=10000,expireAfterWrite=6h, es.metrics.addresses=http://localhost:9200, topology.disruptor.wait.timeout.millis=1000, es.indexer.concurrentRequests=1, robots.noFollow.strict=true, indexer.md.mapping=["parse.title=title" "parse.keywords=keywords" "parse.description=description" "domain=domain"], storm.nimbus.retry.intervalceiling.millis=60000, topology.trident.batch.emit.interval.millis=50, es.status.recentDate.increase=-1, https.protocol.implementation=com.digitalpebble.stormcrawler.protocol.httpclient.HttpProtocol, storm.auth.simple-acl.users=[], drpc.invocations.threads=64, es.status.max.urls.per.bucket=2, java.library.path=/usr/local/lib:/opt/local/lib:/usr/lib, es.indexer.index.name=content, ui.port=8080, fetcher.max.queue.size=-1, selenium.setScriptTimeout=0, indexer.canonical.name=canonical, es.status.bucket.field=key, es.status.routing=true, topology.kryo.decorators=[], es.status.max.buckets=50, storm.id=crawler-1-1574269039, topology.name=crawler, fetcher.max.throttle.sleep=-1, storm.exhibitor.poll.uripath=/exhibitor/v1/cluster/list, fetcher.server.min.delay=0.0, storm.messaging.netty.transfer.batch.size=262144, parsefilters.config.file=parsefilters.json, logviewer.appender.name=A1, nimbus.thrift.max_buffer_size=1048576, storm.auth.simple-acl.users.commands=[], http.protocol.implementation=com.digitalpebble.stormcrawler.protocol.httpclient.HttpProtocol, drpc.request.timeout.secs=600, fetchInterval.error=-1}

6648 [SLOT_1024] INFO  o.a.s.d.worker - Worker f5d45862-968e-43f7-97e8-669e3dae7d11 for storm crawler-1-1574269039 on ab823f98-11b3-4021-b8e7-f9da7ebc12f3:1024 has finished loading

6648 [SLOT_1024] INFO  o.a.s.d.s.Container - SET worker-user f5d45862-968e-43f7-97e8-669e3dae7d11 ubuntu

6649 [SLOT_1024] INFO  o.a.s.d.s.Slot - STATE WAITING_FOR_BLOB_LOCALIZATION msInState: 845 -> WAITING_FOR_WORKER_START msInState: 0 topo:crawler-1-1574269039 worker:f5d45862-968e-43f7-97e8-669e3dae7d11

6649 [SLOT_1024] INFO  o.a.s.d.s.Slot - SLOT 1024: Changing current assignment from null to LocalAssignment(topology_id:crawler-1-1574269039, executors:[ExecutorInfo(task_start:8, task_end:8), ExecutorInfo(task_start:12, task_end:12), ExecutorInfo(task_start:2, task_end:2), ExecutorInfo(task_start:18, task_end:18), ExecutorInfo(task_start:6, task_end:6), ExecutorInfo(task_start:16, task_end:16), ExecutorInfo(task_start:10, task_end:10), ExecutorInfo(task_start:14, task_end:14), ExecutorInfo(task_start:4, task_end:4), ExecutorInfo(task_start:9, task_end:9), ExecutorInfo(task_start:7, task_end:7), ExecutorInfo(task_start:3, task_end:3), ExecutorInfo(task_start:1, task_end:1), ExecutorInfo(task_start:11, task_end:11), ExecutorInfo(task_start:13, task_end:13), ExecutorInfo(task_start:19, task_end:19), ExecutorInfo(task_start:5, task_end:5), ExecutorInfo(task_start:15, task_end:15), ExecutorInfo(task_start:17, task_end:17)], resources:WorkerResources(mem_on_heap:0.0, mem_off_heap:0.0, cpu:0.0), owner:ubuntu)

6653 [SLOT_1024] INFO  o.a.s.d.s.Slot - STATE WAITING_FOR_WORKER_START msInState: 4 topo:crawler-1-1574269039 worker:f5d45862-968e-43f7-97e8-669e3dae7d11 -> RUNNING msInState: 0 topo:crawler-1-1574269039 worker:f5d45862-968e-43f7-97e8-669e3dae7d11

7040 [refresh-active-timer] INFO  o.a.s.d.worker - All connections are ready for worker ab823f98-11b3-4021-b8e7-f9da7ebc12f3:1024 with id f5d45862-968e-43f7-97e8-669e3dae7d11

7055 [Thread-33-spout-executor[9 9]] INFO  o.a.s.d.executor - Opening spout spout:(9)

7056 [Thread-47-spout-executor[10 10]] INFO  o.a.s.d.executor - Opening spout spout:(10)

7058 [Thread-19-spout-executor[12 12]] INFO  o.a.s.d.executor - Opening spout spout:(12)

7069 [Thread-49-spout-executor[14 14]] INFO  o.a.s.d.executor - Opening spout spout:(14)

7071 [Thread-35-spout-executor[11 11]] INFO  o.a.s.d.executor - Opening spout spout:(11)

7079 [Thread-21-__metricscom.digitalpebble.stormcrawler.elasticsearch.metrics.MetricsConsumer-executor[2 2]] INFO  o.a.s.d.executor - Preparing bolt __metricscom.digitalpebble.stormcrawler.elasticsearch.metrics.MetricsConsumer:(2)

7088 [Thread-23-sitemap-executor[7 7]] INFO  o.a.s.d.executor - Preparing bolt sitemap:(7)

7140 [Thread-25-fetcher-executor[3 3]] INFO  o.a.s.d.executor - Preparing bolt fetcher:(3)

7141 [Thread-27-__acker-executor[1 1]] INFO  o.a.s.d.executor - Preparing bolt __acker:(1)

7143 [Thread-39-spout-executor[16 16]] INFO  o.a.s.d.executor - Opening spout spout:(16)

7143 [Thread-31-partitioner-executor[6 6]] INFO  o.a.s.d.executor - Preparing bolt partitioner:(6)

7142 [Thread-43-status_metrics-executor[19 19]] INFO  o.a.s.d.executor - Preparing bolt status_metrics:(19)

7142 [Thread-55-spout-executor[17 17]] INFO  o.a.s.d.executor - Opening spout spout:(17)

7142 [Thread-17-spout-executor[8 8]] INFO  o.a.s.d.executor - Opening spout spout:(8)

7142 [Thread-45-parse-executor[5 5]] INFO  o.a.s.d.executor - Preparing bolt parse:(5)

7142 [Thread-37-__system-executor[-1 -1]] INFO  o.a.s.d.executor - Preparing bolt __system:(-1)

7142 [Thread-53-spout-executor[15 15]] INFO  o.a.s.d.executor - Opening spout spout:(15)

7142 [Thread-41-spout-executor[13 13]] INFO  o.a.s.d.executor - Opening spout spout:(13)

7142 [Thread-29-status-executor[18 18]] INFO  o.a.s.d.executor - Preparing bolt status:(18)

7141 [Thread-51-index-executor[4 4]] INFO  o.a.s.d.executor - Preparing bolt index:(4)

7150 [Thread-37-__system-executor[-1 -1]] INFO  o.a.s.d.executor - Prepared bolt __system:(-1)

7151 [Thread-27-__acker-executor[1 1]] INFO  o.a.s.d.executor - Prepared bolt __acker:(1)

7159 [Thread-31-partitioner-executor[6 6]] INFO  c.d.s.b.URLPartitionerBolt - Using partition mode : byHost

7168 [Thread-31-partitioner-executor[6 6]] INFO  o.a.s.d.executor - Prepared bolt partitioner:(6)

7258 [Thread-29-status-executor[18 18]] INFO  c.d.s.u.URLPartitioner - Using partition mode : byHost

7457 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup BasicURLFilter[com.digitalpebble.stormcrawler.filtering.basic.BasicURLFilter]

7457 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.p.ParseFilters - Setup BasicURLFilter[com.digitalpebble.stormcrawler.filtering.basic.BasicURLFilter]

7458 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup MaxDepthFilter[com.digitalpebble.stormcrawler.filtering.depth.MaxDepthFilter]

7458 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.p.ParseFilters - Setup MaxDepthFilter[com.digitalpebble.stormcrawler.filtering.depth.MaxDepthFilter]

7459 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup BasicURLFilter[com.digitalpebble.stormcrawler.filtering.basic.BasicURLFilter]

7460 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup MaxDepthFilter[com.digitalpebble.stormcrawler.filtering.depth.MaxDepthFilter]

7460 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup BasicURLNormalizer[com.digitalpebble.stormcrawler.filtering.basic.BasicURLNormalizer]

7460 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup BasicURLNormalizer[com.digitalpebble.stormcrawler.filtering.basic.BasicURLNormalizer]

7460 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup HostURLFilter[com.digitalpebble.stormcrawler.filtering.host.HostURLFilter]

7461 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup HostURLFilter[com.digitalpebble.stormcrawler.filtering.host.HostURLFilter]

7462 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.p.ParseFilters - Setup BasicURLNormalizer[com.digitalpebble.stormcrawler.filtering.basic.BasicURLNormalizer]

7462 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.p.ParseFilters - Setup HostURLFilter[com.digitalpebble.stormcrawler.filtering.host.HostURLFilter]

7466 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.p.ParseFilters - Setup RegexURLNormalizer[com.digitalpebble.stormcrawler.filtering.regex.RegexURLNormalizer]

7467 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup RegexURLNormalizer[com.digitalpebble.stormcrawler.filtering.regex.RegexURLNormalizer]

7467 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup RegexURLNormalizer[com.digitalpebble.stormcrawler.filtering.regex.RegexURLNormalizer]

7470 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup RegexURLFilter[com.digitalpebble.stormcrawler.filtering.regex.RegexURLFilter]

7471 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup SelfURLFilter[com.digitalpebble.stormcrawler.filtering.basic.SelfURLFilter]

7473 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.p.ParseFilters - Setup RegexURLFilter[com.digitalpebble.stormcrawler.filtering.regex.RegexURLFilter]

7474 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.p.ParseFilters - Setup SelfURLFilter[com.digitalpebble.stormcrawler.filtering.basic.SelfURLFilter]

7474 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.p.ParseFilters - Setup SitemapFilter[com.digitalpebble.stormcrawler.filtering.sitemap.SitemapFilter]

7473 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup RegexURLFilter[com.digitalpebble.stormcrawler.filtering.regex.RegexURLFilter]

7475 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup SelfURLFilter[com.digitalpebble.stormcrawler.filtering.basic.SelfURLFilter]

7476 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup SitemapFilter[com.digitalpebble.stormcrawler.filtering.sitemap.SitemapFilter]

7473 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup SitemapFilter[com.digitalpebble.stormcrawler.filtering.sitemap.SitemapFilter]

7482 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.b.FetcherBolt - [Fetcher #-1] : starting at 2019-11-20 16:57:22

7520 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup XPathFilter[com.digitalpebble.stormcrawler.parse.filter.XPathFilter]

7532 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup BasicURLFilter[com.digitalpebble.stormcrawler.filtering.basic.BasicURLFilter]

7532 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup MaxDepthFilter[com.digitalpebble.stormcrawler.filtering.depth.MaxDepthFilter]

7532 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup BasicURLNormalizer[com.digitalpebble.stormcrawler.filtering.basic.BasicURLNormalizer]

7532 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup HostURLFilter[com.digitalpebble.stormcrawler.filtering.host.HostURLFilter]

7539 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup RegexURLNormalizer[com.digitalpebble.stormcrawler.filtering.regex.RegexURLNormalizer]

7539 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup XPathFilter[com.digitalpebble.stormcrawler.parse.filter.XPathFilter]

7540 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup RegexURLFilter[com.digitalpebble.stormcrawler.filtering.regex.RegexURLFilter]

7540 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup SelfURLFilter[com.digitalpebble.stormcrawler.filtering.basic.SelfURLFilter]

7540 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup SitemapFilter[com.digitalpebble.stormcrawler.filtering.sitemap.SitemapFilter]

7540 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup LinkParseFilter[com.digitalpebble.stormcrawler.parse.filter.LinkParseFilter]

7540 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup BasicURLFilter[com.digitalpebble.stormcrawler.filtering.basic.BasicURLFilter]

7541 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup MaxDepthFilter[com.digitalpebble.stormcrawler.filtering.depth.MaxDepthFilter]

7541 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup BasicURLNormalizer[com.digitalpebble.stormcrawler.filtering.basic.BasicURLNormalizer]

7541 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup HostURLFilter[com.digitalpebble.stormcrawler.filtering.host.HostURLFilter]

7542 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup RegexURLNormalizer[com.digitalpebble.stormcrawler.filtering.regex.RegexURLNormalizer]

7543 [Thread-45-parse-executor[5 5]] INFO  c.d.s.u.URLPartitioner - Using partition mode : byDomain

7543 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup DomainParseFilter[com.digitalpebble.stormcrawler.parse.filter.DomainParseFilter]

7544 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup RegexURLFilter[com.digitalpebble.stormcrawler.filtering.regex.RegexURLFilter]

7544 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup SelfURLFilter[com.digitalpebble.stormcrawler.filtering.basic.SelfURLFilter]

7544 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup SitemapFilter[com.digitalpebble.stormcrawler.filtering.sitemap.SitemapFilter]

7544 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup LinkParseFilter[com.digitalpebble.stormcrawler.parse.filter.LinkParseFilter]

7544 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.u.URLPartitioner - Using partition mode : byDomain

7544 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup DomainParseFilter[com.digitalpebble.stormcrawler.parse.filter.DomainParseFilter]

7545 [Thread-23-sitemap-executor[7 7]] INFO  c.d.s.p.ParseFilters - Setup CommaSeparatedToMultivaluedMetadata[com.digitalpebble.stormcrawler.parse.filter.CommaSeparatedToMultivaluedMetadata]

7545 [Thread-45-parse-executor[5 5]] INFO  c.d.s.p.ParseFilters - Setup CommaSeparatedToMultivaluedMetadata[com.digitalpebble.stormcrawler.parse.filter.CommaSeparatedToMultivaluedMetadata]

7550 [Thread-45-parse-executor[5 5]] INFO  o.a.s.d.executor - Prepared bolt parse:(5)

7550 [Thread-23-sitemap-executor[7 7]] INFO  o.a.s.d.executor - Prepared bolt sitemap:(7)

7718 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.p.RobotRulesParser - No agents listed in 'http.robots.agents' property! Using http.agent.name [Google-Read-Aloud]

7722 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.p.RobotRulesParser - No agents listed in 'http.robots.agents' property! Using http.agent.name [Google-Read-Aloud]

7722 [Thread-25-fetcher-executor[3 3]] INFO  c.d.s.b.FetcherBolt - Using queue mode : byHost

7792 [Thread-25-fetcher-executor[3 3]] INFO  o.a.s.d.executor - Prepared bolt fetcher:(3)

7820 [Thread-39-spout-executor[16 16]] INFO  c.d.s.e.p.AbstractSpout - [spout #8]  assigned shard ID 8

7823 [Thread-55-spout-executor[17 17]] INFO  c.d.s.e.p.AbstractSpout - [spout #9]  assigned shard ID 9

7824 [Thread-17-spout-executor[8 8]] INFO  c.d.s.e.p.AbstractSpout - [spout #0]  assigned shard ID 0

7824 [Thread-41-spout-executor[13 13]] INFO  c.d.s.e.p.AbstractSpout - [spout #5]  assigned shard ID 5

7825 [Thread-53-spout-executor[15 15]] INFO  c.d.s.e.p.AbstractSpout - [spout #7]  assigned shard ID 7

7825 [Thread-35-spout-executor[11 11]] INFO  c.d.s.e.p.AbstractSpout - [spout #3]  assigned shard ID 3

7825 [Thread-35-spout-executor[11 11]] ERROR o.a.s.util - Async loop died!

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7821 [Thread-39-spout-executor[16 16]] ERROR o.a.s.util - Async loop died!

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7828 [Thread-35-spout-executor[11 11]] ERROR o.a.s.d.executor - 

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7828 [Thread-39-spout-executor[16 16]] ERROR o.a.s.d.executor - 

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7825 [Thread-47-spout-executor[10 10]] INFO  c.d.s.e.p.AbstractSpout - [spout #2]  assigned shard ID 2

7829 [Thread-47-spout-executor[10 10]] ERROR o.a.s.util - Async loop died!

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7829 [Thread-47-spout-executor[10 10]] ERROR o.a.s.d.executor - 

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7825 [Thread-19-spout-executor[12 12]] INFO  c.d.s.e.p.AbstractSpout - [spout #4]  assigned shard ID 4

7830 [Thread-19-spout-executor[12 12]] ERROR o.a.s.util - Async loop died!

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7825 [Thread-53-spout-executor[15 15]] ERROR o.a.s.util - Async loop died!

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7830 [Thread-19-spout-executor[12 12]] ERROR o.a.s.d.executor - 

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7830 [Thread-53-spout-executor[15 15]] ERROR o.a.s.d.executor - 

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7824 [Thread-41-spout-executor[13 13]] ERROR o.a.s.util - Async loop died!

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7831 [Thread-41-spout-executor[13 13]] ERROR o.a.s.d.executor - 

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7824 [Thread-17-spout-executor[8 8]] ERROR o.a.s.util - Async loop died!

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7840 [Thread-17-spout-executor[8 8]] ERROR o.a.s.d.executor - 

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7823 [Thread-55-spout-executor[17 17]] ERROR o.a.s.util - Async loop died!

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7825 [Thread-49-spout-executor[14 14]] INFO  c.d.s.e.p.AbstractSpout - [spout #6]  assigned shard ID 6

7842 [Thread-49-spout-executor[14 14]] ERROR o.a.s.util - Async loop died!

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7841 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f2000f type:create cxid:0xdc zxid:0x3c txntype:-1 reqpath:n/a Error Path:/storm/errors/crawler-1-1574269039/spout Error:KeeperErrorCode = NodeExists for /storm/errors/crawler-1-1574269039/spout

7842 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f2000f type:create cxid:0xdf zxid:0x3d txntype:-1 reqpath:n/a Error Path:/storm/errors/crawler-1-1574269039/spout Error:KeeperErrorCode = NodeExists for /storm/errors/crawler-1-1574269039/spout

7842 [Thread-49-spout-executor[14 14]] ERROR o.a.s.d.executor - 

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7843 [Thread-55-spout-executor[17 17]] ERROR o.a.s.d.executor - 

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7845 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f2000f type:create cxid:0xe6 zxid:0x3e txntype:-1 reqpath:n/a Error Path:/storm/errors/crawler-1-1574269039/spout Error:KeeperErrorCode = NodeExists for /storm/errors/crawler-1-1574269039/spout

7839 [Thread-33-spout-executor[9 9]] INFO  c.d.s.e.p.AbstractSpout - [spout #1]  assigned shard ID 1

7846 [Thread-33-spout-executor[9 9]] ERROR o.a.s.util - Async loop died!

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7846 [Thread-33-spout-executor[9 9]] ERROR o.a.s.d.executor - 

java.lang.ClassCastException: class clojure.lang.PersistentVector cannot be cast to class java.lang.String (clojure.lang.PersistentVector is in unnamed module of loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')

        at com.digitalpebble.stormcrawler.util.ConfUtils.getString(ConfUtils.java:74) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AbstractSpout.open(AbstractSpout.java:188) ~[dev1-0.1.jar:?]

        at com.digitalpebble.stormcrawler.elasticsearch.persistence.AggregationSpout.open(AggregationSpout.java:98) ~[dev1-0.1.jar:?]

        at org.apache.storm.daemon.executor$fn__10112$fn__10127.invoke(executor.clj:609) ~[storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:482) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7847 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f2000f type:create cxid:0xf0 zxid:0x42 txntype:-1 reqpath:n/a Error Path:/storm/errors/crawler-1-1574269039/spout Error:KeeperErrorCode = NodeExists for /storm/errors/crawler-1-1574269039/spout

7863 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f2000f type:create cxid:0x105 zxid:0x4b txntype:-1 reqpath:n/a Error Path:/storm/errors/crawler-1-1574269039/spout-last-error Error:KeeperErrorCode = NodeExists for /storm/errors/crawler-1-1574269039/spout-last-error

7863 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f2000f type:create cxid:0x106 zxid:0x4c txntype:-1 reqpath:n/a Error Path:/storm/errors/crawler-1-1574269039/spout-last-error Error:KeeperErrorCode = NodeExists for /storm/errors/crawler-1-1574269039/spout-last-error

7875 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f2000f type:create cxid:0x111 zxid:0x50 txntype:-1 reqpath:n/a Error Path:/storm/errors/crawler-1-1574269039/spout-last-error Error:KeeperErrorCode = NodeExists for /storm/errors/crawler-1-1574269039/spout-last-error

7875 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f2000f type:create cxid:0x112 zxid:0x51 txntype:-1 reqpath:n/a Error Path:/storm/errors/crawler-1-1574269039/spout-last-error Error:KeeperErrorCode = NodeExists for /storm/errors/crawler-1-1574269039/spout-last-error

7875 [ProcessThread(sid:0 cport:2000):] INFO  o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x100040091f2000f type:create cxid:0x114 zxid:0x53 txntype:-1 reqpath:n/a Error Path:/storm/errors/crawler-1-1574269039/spout-last-error Error:KeeperErrorCode = NodeExists for /storm/errors/crawler-1-1574269039/spout-last-error

7886 [Thread-19-spout-executor[12 12]] ERROR o.a.s.util - Halting process: ("Worker died")

java.lang.RuntimeException: ("Worker died")

        at org.apache.storm.util$exit_process_BANG_.doInvoke(util.clj:341) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.RestFn.invoke(RestFn.java:423) [clojure-1.7.0.jar:?]

        at org.apache.storm.daemon.worker$fn__10799$fn__10800.invoke(worker.clj:788) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.daemon.executor$mk_executor_data$fn__9997$fn__9998.invoke(executor.clj:281) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:494) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7886 [Thread-17-spout-executor[8 8]] ERROR o.a.s.util - Halting process: ("Worker died")

java.lang.RuntimeException: ("Worker died")

        at org.apache.storm.util$exit_process_BANG_.doInvoke(util.clj:341) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.RestFn.invoke(RestFn.java:423) [clojure-1.7.0.jar:?]

        at org.apache.storm.daemon.worker$fn__10799$fn__10800.invoke(worker.clj:788) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.daemon.executor$mk_executor_data$fn__9997$fn__9998.invoke(executor.clj:281) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:494) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7886 [Thread-49-spout-executor[14 14]] ERROR o.a.s.util - Halting process: ("Worker died")

java.lang.RuntimeException: ("Worker died")

        at org.apache.storm.util$exit_process_BANG_.doInvoke(util.clj:341) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.RestFn.invoke(RestFn.java:423) [clojure-1.7.0.jar:?]

        at org.apache.storm.daemon.worker$fn__10799$fn__10800.invoke(worker.clj:788) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.daemon.executor$mk_executor_data$fn__9997$fn__9998.invoke(executor.clj:281) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:494) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7892 [Thread-39-spout-executor[16 16]] ERROR o.a.s.util - Halting process: ("Worker died")

java.lang.RuntimeException: ("Worker died")

        at org.apache.storm.util$exit_process_BANG_.doInvoke(util.clj:341) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.RestFn.invoke(RestFn.java:423) [clojure-1.7.0.jar:?]

        at org.apache.storm.daemon.worker$fn__10799$fn__10800.invoke(worker.clj:788) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.daemon.executor$mk_executor_data$fn__9997$fn__9998.invoke(executor.clj:281) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:494) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7892 [Thread-35-spout-executor[11 11]] ERROR o.a.s.util - Halting process: ("Worker died")

java.lang.RuntimeException: ("Worker died")

        at org.apache.storm.util$exit_process_BANG_.doInvoke(util.clj:341) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.RestFn.invoke(RestFn.java:423) [clojure-1.7.0.jar:?]

        at org.apache.storm.daemon.worker$fn__10799$fn__10800.invoke(worker.clj:788) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.daemon.executor$mk_executor_data$fn__9997$fn__9998.invoke(executor.clj:281) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:494) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7893 [Thread-47-spout-executor[10 10]] ERROR o.a.s.util - Halting process: ("Worker died")

java.lang.RuntimeException: ("Worker died")

        at org.apache.storm.util$exit_process_BANG_.doInvoke(util.clj:341) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.RestFn.invoke(RestFn.java:423) [clojure-1.7.0.jar:?]

        at org.apache.storm.daemon.worker$fn__10799$fn__10800.invoke(worker.clj:788) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.daemon.executor$mk_executor_data$fn__9997$fn__9998.invoke(executor.clj:281) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:494) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7895 [Thread-55-spout-executor[17 17]] ERROR o.a.s.util - Halting process: ("Worker died")

java.lang.RuntimeException: ("Worker died")

        at org.apache.storm.util$exit_process_BANG_.doInvoke(util.clj:341) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.RestFn.invoke(RestFn.java:423) [clojure-1.7.0.jar:?]

        at org.apache.storm.daemon.worker$fn__10799$fn__10800.invoke(worker.clj:788) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.daemon.executor$mk_executor_data$fn__9997$fn__9998.invoke(executor.clj:281) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:494) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7895 [Thread-41-spout-executor[13 13]] ERROR o.a.s.util - Halting process: ("Worker died")

java.lang.RuntimeException: ("Worker died")

        at org.apache.storm.util$exit_process_BANG_.doInvoke(util.clj:341) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.RestFn.invoke(RestFn.java:423) [clojure-1.7.0.jar:?]

        at org.apache.storm.daemon.worker$fn__10799$fn__10800.invoke(worker.clj:788) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.daemon.executor$mk_executor_data$fn__9997$fn__9998.invoke(executor.clj:281) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:494) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

7895 [Thread-53-spout-executor[15 15]] ERROR o.a.s.util - Halting process: ("Worker died")

java.lang.RuntimeException: ("Worker died")

        at org.apache.storm.util$exit_process_BANG_.doInvoke(util.clj:341) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.RestFn.invoke(RestFn.java:423) [clojure-1.7.0.jar:?]

        at org.apache.storm.daemon.worker$fn__10799$fn__10800.invoke(worker.clj:788) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.daemon.executor$mk_executor_data$fn__9997$fn__9998.invoke(executor.clj:281) [storm-core-1.2.3.jar:1.2.3]

        at org.apache.storm.util$async_loop$fn__624.invoke(util.clj:494) [storm-core-1.2.3.jar:1.2.3]

        at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]

        at java.lang.Thread.run(Thread.java:834) [?:?]

DigitalPebble

unread,
Nov 20, 2019, 3:12:52 PM11/20/19
to DigitalPebble

--
You received this message because you are subscribed to the Google Groups "DigitalPebble" group.
To unsubscribe from this group and stop receiving emails from it, send an email to digitalpebbl...@googlegroups.com.
To view this discussion on the web, visit https://groups.google.com/d/msgid/digitalpebble/5e39215b-c3b5-4d96-90b0-a6c6391ab0fa%40googlegroups.com.


--
Reply all
Reply to author
Forward
0 new messages