druid.indexer.runner.javaOpts=-server -Xmx3g -XX:+UseG1GC -XX:MaxGCPauseMillis=100 -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -Duser.timezone=UTC -Dfile.encoding=UTF-8 -Djava.util.logging.manager=org.apache.logging.log4j.jul.LogManager -Dhadoop.mapreduce.job.user.classpath.first=true
My rows have 132 columns, but even conservatively guesstimating that they require 128 bytes each (they require less) that would still come in at a little over 800MB for 50k rows which doesn't seem excessive given 3GB configured for the jvm.
Am I missing a setting? Bad math somewhere?
Thanks,
--Ben
--
You received this message because you are subscribed to the Google Groups "Druid User" group.
To unsubscribe from this group and stop receiving emails from it, send an email to druid-user+...@googlegroups.com.
To post to this group, send email to druid...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/druid-user/f9f57182-62e4-4450-be53-7ccdbf4eb553%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
2016-07-21T02:38:50,750 ERROR [task-runner-0-priority-0] io.druid.indexing.overlord.ThreadPoolTaskRunner - Exception while running task[KafkaIndexTask{id=index_kafka_business_events_test_191cda0fb891da5_nmiakddb, type=index_kafka, dataSource=business_events_test}]
java.lang.RuntimeException: java.lang.InterruptedException
at com.google.common.base.Throwables.propagate(Throwables.java:160) ~[guava-16.0.1.jar:?]
at io.druid.indexing.common.actions.RemoteTaskActionClient.submit(RemoteTaskActionClient.java:110) ~[druid-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.appenderator.ActionBasedSegmentAllocator.allocate(ActionBasedSegmentAllocator.java:52) ~[druid-indexing-service-0.9.1.jar:0.9.1]
at io.druid.segment.realtime.appenderator.FiniteAppenderatorDriver.getSegment(FiniteAppenderatorDriver.java:341) ~[druid-server-0.9.1.jar:0.9.1]
at io.druid.segment.realtime.appenderator.FiniteAppenderatorDriver.add(FiniteAppenderatorDriver.java:195) ~[druid-server-0.9.1.jar:0.9.1]
at io.druid.indexing.kafka.KafkaIndexTask.run(KafkaIndexTask.java:417) ~[?:?]
at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:436) [druid-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:408) [druid-indexing-service-0.9.1.jar:0.9.1]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_91]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_91]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_91]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_91]
Caused by: java.lang.InterruptedException
at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1302) ~[?:1.8.0_91]
at com.google.common.util.concurrent.AbstractFuture$Sync.get(AbstractFuture.java:285) ~[guava-16.0.1.jar:?]
at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:116) ~[guava-16.0.1.jar:?]
at io.druid.indexing.common.actions.RemoteTaskActionClient.submit(RemoteTaskActionClient.java:105) ~[druid-indexing-service-0.9.1.jar:0.9.1]
... 10 more
2016-07-21T02:38:50,772 INFO [task-runner-0-priority-0] io.druid.indexing.overlord.TaskRunnerUtils - Task [index_kafka_business_events_test_191cda0fb891da5_nmiakddb] status changed to [FAILED].
2016-07-21T02:38:50,779 INFO [task-runner-0-priority-0] io.druid.indexing.worker.executor.ExecutorLifecycle - Task completed with status: {
"id" : "index_kafka_business_events_test_191cda0fb891da5_nmiakddb",
"status" : "FAILED",
"duration" : 1954254
}
Also, despite having only 2 indexing tasks running (according to the web console) I still have a whole bunch of peon processes:
[hdfs@druid-data002 services]$ jcmd
55777 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_business_events_test_191cda0fb891da5_njpngnnb/task.json /services/druid/data2/task/index_kafka_business_events_test_191cda0fb891da5_njpngnnb/4db94321-10c2-4165-9c2c-fb30ee7200a0/status.json
48290 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_sor_business_event_success_v1_bc9136377a73b79_lippeobc/task.json /services/druid/data2/task/index_kafka_sor_business_event_success_v1_bc9136377a73b79_lippeobc/c1d6e5aa-bd2c-4d28-82e9-1596a5769946/status.json
47524 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_business_events_test_dbdc87080a0fca8_phgdbnbn/task.json /services/druid/data2/task/index_kafka_business_events_test_dbdc87080a0fca8_phgdbnbn/f03b47d3-cdb4-4b61-b616-c7e9e1ed9542/status.json
51653 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_sor_business_event_success_v1_ceccd9ff870f2b1_ekhfhomh/task.json /services/druid/data2/task/index_kafka_sor_business_event_success_v1_ceccd9ff870f2b1_ekhfhomh/e95e4dfc-ce8b-48f5-92ce-29f89cb940ce/status.json
58508 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_sor_business_event_success_v1_6326aca7c5a400e_khnbjkig/task.json /services/druid/data2/task/index_kafka_sor_business_event_success_v1_6326aca7c5a400e_khnbjkig/8be4cf00-77bb-4f8d-ab76-a05ca04d4608/status.json
28562 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_sor_business_event_success_v1_3bc7e5362bc9b22_odfhmkal/task.json /services/druid/data2/task/index_kafka_sor_business_event_success_v1_3bc7e5362bc9b22_odfhmkal/2fc0f8a7-e1a0-480a-ab79-c30edaeb8052/status.json
16343 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_sor_business_event_success_v1_5b8922454a6ba85_djjhelld/task.json /services/druid/data2/task/index_kafka_sor_business_event_success_v1_5b8922454a6ba85_djjhelld/07da9147-2f39-4d53-85f6-8512c62c9796/status.json
25851 io.druid.cli.Main server historical
25852 io.druid.cli.Main server middleManager
61021 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_sor_business_event_success_v1_590eacdd7bca8fb_gjpebjpc/task.json /services/druid/data2/task/index_kafka_sor_business_event_success_v1_590eacdd7bca8fb_gjpebjpc/24ad0c3e-525f-401d-8db2-a0fbea6ee3ad/status.json
39965 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_sor_business_event_success_v1_1795277419cc5ef_jbbdoalg/task.json /services/druid/data2/task/index_kafka_sor_business_event_success_v1_1795277419cc5ef_jbbdoalg/19a1a0cc-920f-448c-83d5-60ce32e98035/status.json
57790 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_business_events_test_bb8b4037b1ecca7_dihbeaff/task.json /services/druid/data2/task/index_kafka_business_events_test_bb8b4037b1ecca7_dihbeaff/0e205d2b-43a0-4d46-9333-0035be34eef1/status.json
41182 io.druid.cli.Main internal peon /services/druid/data2/task/index_kafka_business_events_test_a6e41386ef80784_obdopepl/task.json /services/druid/data2/task/index_kafka_business_events_test_a6e41386ef80784_obdopepl/a73a7ccd-2913-4302-8529-1246137c769d/status.json
2016-07-21T04:44:07,964 ERROR [task-runner-0-priority-0] io.druid.indexing.overlord.ThreadPoolTaskRunner - Exception while running task[KafkaIndexTask{id=index_kafka_business_events_test_ad1379bb84a7460_pgeigkmd, type=index_kafka, dataSource=business_events_test}]
com.metamx.common.ISE: Missing hydrant [2] in sinkDir [/services/druid/data2/task/index_kafka_business_events_test_ad1379bb84a7460_pgeigkmd/work/persist/business_events_test_2016-07-15T15:00:00.000Z_2016-07-15T16:00:00.000Z_2016-07-21T03:34:14.236Z_13].
at io.druid.segment.realtime.appenderator.AppenderatorImpl.bootstrapSinksFromDisk(AppenderatorImpl.java:960) ~[druid-server-0.9.1.jar:0.9.1]
at io.druid.segment.realtime.appenderator.AppenderatorImpl.startJob(AppenderatorImpl.java:197) ~[druid-server-0.9.1.jar:0.9.1]
at io.druid.segment.realtime.appenderator.FiniteAppenderatorDriver.startJob(FiniteAppenderatorDriver.java:136) ~[druid-server-0.9.1.jar:0.9.1]
at io.druid.indexing.kafka.KafkaIndexTask.run(KafkaIndexTask.java:275) ~[?:?]
at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:436) [druid-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:408) [druid-indexing-service-0.9.1.jar:0.9.1]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_91]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_91]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_91]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_91]
Suppressed: java.lang.NullPointerException
at io.druid.segment.realtime.appenderator.AppenderatorImpl.mergeBarrier(AppenderatorImpl.java:652) ~[druid-server-0.9.1.jar:0.9.1]
at io.druid.segment.realtime.appenderator.AppenderatorImpl.abandonSegment(AppenderatorImpl.java:1021) ~[druid-server-0.9.1.jar:0.9.1]
at io.druid.segment.realtime.appenderator.AppenderatorImpl.close(AppenderatorImpl.java:758) ~[druid-server-0.9.1.jar:0.9.1]
at io.druid.indexing.kafka.KafkaIndexTask.run(KafkaIndexTask.java:526) ~[?:?]
at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:436) [druid-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:408) [druid-indexing-service-0.9.1.jar:0.9.1]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_91]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_91]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_91]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_91]
2016-07-21T04:44:07,975 INFO [task-runner-0-priority-0] io.druid.indexing.overlord.TaskRunnerUtils - Task [index_kafka_business_events_test_ad1379bb84a7460_pgeigkmd] status changed to [FAILED].
2016-07-21T04:44:07,980 INFO [task-runner-0-priority-0] io.druid.indexing.worker.executor.ExecutorLifecycle - Task completed with status: {
"id" : "index_kafka_business_events_test_ad1379bb84a7460_pgeigkmd",
"status" : "FAILED",
"duration" : 3297
}
And I found a couple of these in the overlord log:
2016-07-21T04:31:44,986 INFO [qtp690051320-215] io.druid.indexing.overlord.TaskLockbox - Task[index_kafka_business_events_test_ad1379bb84a7460_pgeigkmd] already present in TaskLock[index_kafka_business_events_test]
2016-07-21T04:31:44,998 INFO [qtp690051320-215] io.druid.metadata.IndexerSQLMetadataStorageCoordinator - Allocated pending segment [business_events_test_2016-07-14T11:00:00.000Z_2016-07-14T12:00:00.000Z_2016-07-21T03:30:30.914Z_9] for sequence[index_kafka_business_events_test_ad1379bb84a7460_5] (previous = [business_events_test_2016-07-14T10:00:00.000Z_2016-07-14T11:00:00.000Z_2016-07-21T03:30:28.742Z_9]) in DB
2016-07-21T04:31:45,093 WARN [KafkaSupervisor-sor_business_event_success_v1-0] io.druid.indexing.kafka.supervisor.KafkaSupervisor - Task [index_kafka_sor_business_event_success_v1_feffecbde70f303_bcjmfelf] failed to return start time, killing task
java.lang.RuntimeException: java.net.ConnectException: Connection refused
at com.google.common.base.Throwables.propagate(Throwables.java:160) ~[guava-16.0.1.jar:?]
at io.druid.indexing.kafka.KafkaIndexTaskClient.submitRequest(KafkaIndexTaskClient.java:328) ~[druid-kafka-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.kafka.KafkaIndexTaskClient.submitRequest(KafkaIndexTaskClient.java:233) ~[druid-kafka-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.kafka.KafkaIndexTaskClient.getStartTime(KafkaIndexTaskClient.java:156) ~[druid-kafka-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.kafka.supervisor.KafkaSupervisor.getTaskStartTime(KafkaSupervisor.java:1218) ~[druid-kafka-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.kafka.supervisor.KafkaSupervisor.updateTaskStatus(KafkaSupervisor.java:708) [druid-kafka-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.kafka.supervisor.KafkaSupervisor.runInternal(KafkaSupervisor.java:457) [druid-kafka-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.kafka.supervisor.KafkaSupervisor$RunNotice.handle(KafkaSupervisor.java:409) [druid-kafka-indexing-service-0.9.1.jar:0.9.1]
at io.druid.indexing.kafka.supervisor.KafkaSupervisor$2.run(KafkaSupervisor.java:261) [druid-kafka-indexing-service-0.9.1.jar:0.9.1]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [?:1.8.0_91]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_91]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_91]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_91]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_91]
Caused by: java.net.ConnectException: Connection refused
at java.net.PlainSocketImpl.socketConnect(Native Method) ~[?:1.8.0_91]
at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350) ~[?:1.8.0_91]
at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206) ~[?:1.8.0_91]
at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188) ~[?:1.8.0_91]
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) ~[?:1.8.0_91]
at java.net.Socket.connect(Socket.java:589) ~[?:1.8.0_91]
at java.net.Socket.connect(Socket.java:538) ~[?:1.8.0_91]
at java.net.Socket.<init>(Socket.java:434) ~[?:1.8.0_91]
at java.net.Socket.<init>(Socket.java:211) ~[?:1.8.0_91]
at io.druid.indexing.kafka.KafkaIndexTaskClient.checkConnection(KafkaIndexTaskClient.java:228) ~[?:?]
at io.druid.indexing.kafka.KafkaIndexTaskClient.submitRequest(KafkaIndexTaskClient.java:264) ~[?:?]
... 12 more
SET GLOBAL max_allowed_packet = 1024*1024*16;
Maybe give that a try and see if it helps at all?
--
You received this message because you are subscribed to the Google Groups "Druid User" group.
To unsubscribe from this group and stop receiving emails from it, send an email to druid-user+...@googlegroups.com.
To post to this group, send email to druid...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/druid-user/466aa9c0-32ca-4fe3-98ae-9b1feba82193%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.