--
You received this message because you are subscribed to the Google Groups "Druid Development" group.
To unsubscribe from this group and stop receiving emails from it, send an email to druid-developm...@googlegroups.com.
To post to this group, send email to druid-de...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/druid-development/b24d2b02-3c88-49d3-8359-c2a1ea884810%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
# Change these to make Druid faster
druid.processing.buffer.sizeBytes=100000000
druid.processing.numThreads=1
druid.host=localhost
druid.service=broker
druid.port=8084
druid.zk.service.host=titan3,titan2,titan1
druid.db.connector.connectURI=jdbc:mysql://db1:3306/druid
druid.db.connector.user=druid
druid.db.connector.password=druid
druid.coordinator.startDelay=PT70s
druid.indexer.logs.directory=/tmp/druid/log
druid.request.logging.dir=/tmp/druid/log
com.metamx.emitter.logging.level=error
druid.emitter.logging.logLevel=error
druid.host=localhost
druid.service=historical
druid.port=8081
druid.zk.service.host=titan3,titan2,titan1
druid.extensions.coordinates=["io.druid.extensions:druid-hdfs-storage:0.6.147"]
druid.server.maxSize=10000000000
# Change these to make Druid faster
druid.processing.buffer.sizeBytes=100000000
druid.processing.numThreads=1
druid.segmentCache.locations=[{"path": "/tmp/druid/indexCache", "maxSize": 10000000000}]
druid.indexer.logs.directory=/tmp/druid/log
druid.request.logging.dir=/tmp/druid/log
com.metamx.emitter.logging.level=error
druid.emitter.logging.logLevel=error
druid.host=localhost
druid.service=coordinator
druid.port=8082
druid.zk.service.host=titan3,titan2,titan1
druid.db.connector.connectURI=jdbc:mysql://db1:3306/druid
druid.db.connector.user=druid
druid.db.connector.password=druid
druid.coordinator.startDelay=PT70s
druid.indexer.logs.directory=/tmp/druid/log
druid.request.logging.dir=/tmp/druid/log
com.metamx.emitter.logging.level=error
druid.emitter.logging.logLevel=error
To view this discussion on the web visit https://groups.google.com/d/msgid/druid-development/60bbfe22-8b10-4942-9238-b6d502a55fc8%40googlegroups.com.
druid.processing.numThreads=5
druid.processing.buffer.sizeBytes=1142857142
So to extend my POC, i indexed 100K records and i see the following in the coordinator log
Coordinator Logs
2014-12-30 19:32:13,061 INFO [main-EventThread] io.druid.server.coordinator.LoadQueuePeon - Server[/druid/loadQueue/localhost:8083] done processing [/druid/loadQueue/localhost:8083/pci_2012-01-01T00:00:00.000Z_2012-01-02T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:32:33,288 INFO [DatabaseRuleManager-Exec--0] io.druid.db.DatabaseRuleManager - Polled and found rules for 1 datasource(s)
2014-12-30 19:32:33,923 INFO [DatabaseSegmentManager-Exec--0] io.druid.db.DatabaseSegmentManager - Polled and found 735 segments in the database
But when i query broker i see that the query is taking a really long time and see the following in the borker logs (historical runs on 8083, broker runs on 8084)
Historical Logs
2014-12-30 19:34:48,045 INFO [qtp649630909-18] io.druid.query.GroupByParallelQueryRunner - Query timeout, cancelling pending results for query id [7d6984b1-94ed-4457-a68d-9c5362b55687]
Broker Logs
2014-12-30 19:30:43,480 INFO [qtp1374432753-42] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://localhost:8083
2014-12-30 19:31:09,816 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2013-12-24T00:00:00.000Z_2013-12-25T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:10,456 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2013-02-16T00:00:00.000Z_2013-02-17T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:10,488 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2013-02-08T00:00:00.000Z_2013-02-09T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:10,588 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2013-01-01T00:00:00.000Z_2013-01-02T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:10,635 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-12-31T00:00:00.000Z_2013-01-01T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:10,700 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-12-13T00:00:00.000Z_2012-12-14T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:10,747 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-12-05T00:00:00.000Z_2012-12-06T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:10,787 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-11-12T00:00:00.000Z_2012-11-13T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:10,849 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-11-03T00:00:00.000Z_2012-11-04T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:10,948 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-09-24T00:00:00.000Z_2012-09-25T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:10,987 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-08-27T00:00:00.000Z_2012-08-28T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,052 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-08-09T00:00:00.000Z_2012-08-10T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,067 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-07-18T00:00:00.000Z_2012-07-19T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,101 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-06-20T00:00:00.000Z_2012-06-21T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,126 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-06-01T00:00:00.000Z_2012-06-02T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,179 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-05-15T00:00:00.000Z_2012-05-16T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,208 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-04-29T00:00:00.000Z_2012-04-30T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,240 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-03-26T00:00:00.000Z_2012-03-27T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,282 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-02-29T00:00:00.000Z_2012-03-01T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,309 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-02-21T00:00:00.000Z_2012-02-22T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,323 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] removed segment[pci_2012-12-01T00:00:00.000Z_2012-12-02T00:00:00.000Z_2014-12-29T22:43:01.376Z]
2014-12-30 19:31:11,398 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-02-09T00:00:00.000Z_2012-02-10T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,424 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-01-26T00:00:00.000Z_2012-01-27T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:11,466 INFO [ServerInventoryView-0] io.druid.client.SingleServerInventoryView - Server[localhost:8083] added segment[pci_2012-01-10T00:00:00.000Z_2012-01-11T00:00:00.000Z_2014-12-30T19:03:31.590Z]
2014-12-30 19:31:12,540 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.539Z","service":"broker","host":"localhost:8084","metric":"cache/delta/numEntries","value":0}]
2014-12-30 19:31:12,540 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.540Z","service":"broker","host":"localhost:8084","metric":"cache/delta/sizeBytes","value":0}]
2014-12-30 19:31:12,540 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.540Z","service":"broker","host":"localhost:8084","metric":"cache/delta/hits","value":0}]
2014-12-30 19:31:12,540 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.540Z","service":"broker","host":"localhost:8084","metric":"cache/delta/misses","value":0}]
2014-12-30 19:31:12,540 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.540Z","service":"broker","host":"localhost:8084","metric":"cache/delta/evictions","value":0}]
2014-12-30 19:31:12,540 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.540Z","service":"broker","host":"localhost:8084","metric":"cache/delta/hitRate","value":0.0}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/delta/averageBytes","value":0}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/delta/timeouts","value":0}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/delta/errors","value":0}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/total/numEntries","value":0}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/total/sizeBytes","value":0}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/total/hits","value":0}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/total/misses","value":40}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/total/evictions","value":40}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/total/hitRate","value":0.0}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/total/averageBytes","value":0}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/total/timeouts","value":0}]
2014-12-30 19:31:12,541 ERROR [MonitorScheduler-0] com.metamx.emitter.core.LoggingEmitter - Event [{"feed":"metrics","timestamp":"2014-12-30T19:31:12.541Z","service":"broker","host":"localhost:8084","metric":"cache/total/errors","value":0}]
curl -X 'POST' -H 'Content-Type:application/json' -d @index.json http://localhost:8080/druid/indexer/v1/task
The runtime.properties of historical nodes have this setup
druid.extensions.coordinates=["io.druid.extensions:druid-hdfs-storage:0.6.160"]
druid.storage.storageDirectory=hdfs://nn1/data/druid/pci
druid.storage.type=hdfs
...
druid.extensions.coordinates=["io.druid.extensions:druid-hdfs-storage:0.6.160"]
druid.storage.storageDirectory=hdfs://nn1:8020/data/druid/pci
druid.storage.type=hdfs
Indexing logs :
2014-12-31 02:15:31,838 WARN [task-runner-0] io.druid.indexing.common.index.YeOldePlumberSchool - Failed to merge and upload java.io.IOException: No FileSystem for scheme: hdfs at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2304) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2311) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:90) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2350) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2332) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:369) at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296) at io.druid.storage.hdfs.HdfsDataSegmentPusher.push(HdfsDataSegmentPusher.java:75) at io.druid.indexing.common.task.IndexTask$2.push(IndexTask.java:390) at io.druid.indexing.common.index.YeOldePlumberSchool$1.finishJob(YeOldePlumberSchool.java:179) at io.druid.indexing.common.task.IndexTask.generateSegment(IndexTask.java:444) at io.druid.indexing.common.task.IndexTask.run(IndexTask.java:198) at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:218) at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:197) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) 2014-12-31 02:15:31,842 INFO [task-runner-0] io.druid.indexing.common.index.YeOldePlumberSchool - Del
...
...
...
-rw-r--r--  3 sudhrao hdfs    537 2014-12-30 19:42 /data/druid/pci/20120401T000000.000Z_20120501T000000.000Z/2014-12-31T02_41_55.040Z/0/descriptor.json
-rw-r--r--  3 sudhrao hdfs   285372 2014-12-30 19:42 /data/druid/pci/20120401T000000.000Z_20120501T000000.000Z/2014-12-31T02_41_55.040Z/0/index.zip
-rw-r--r--  3 sudhrao hdfs    537 2014-12-30 19:42 /data/druid/pci/20120501T000000.000Z_20120601T000000.000Z/2014-12-31T02_41_55.040Z/0/descriptor.json
-rw-r--r--  3 sudhrao hdfs   244960 2014-12-30 19:42 /data/druid/pci/20120501T000000.000Z_20120601T000000.000Z/2014-12-31T02_41_55.040Z/0/index.zip
...
....
however when i see the coordinator cluster.html page, I see only one server having data. Is this expected due to my data size ? why don;t other historical nodes load the segments from hdfs
PS : checkout the screenshot in the attachement
...
...
...
{
  "type":"index",
  "dataSource" : "pci",
  "granularitySpec" : {
      "type" : "uniform",
      "gran" : "month",
      "intervals" : [ "2000/2016" ]
  },
  "aggregators":[
          {"type":"doubleSum","name":"count","fieldName":"count"},
          {"type":"doubleSum","name":"amount","fieldName":"amount"}
    ],
 "config": {
   "maxRowsInMemory": 500000000,
   "intermediatePersistPeriod": "PT10m"
  },
  "firehose" : {
   "type" : "local",
    "baseDir" : "/home/sudhrao",
   "filter" : "100k.json",
   "parser" : {
   "timestampSpec" : {
    "column" : "timestamp"
   },
   "data": {
    "format": "json",
    "columns": ["timestamp","account_number","tcode","amount","currency","country","timezone","count"],
    "dimensions": ["account_number","tcode","curreny","country","timezone"]
   }
  }
 },
 "plumber": {
   "type": "realtime",
   "windowPeriod": "PT10m",
   "segmentGranularity": "day",
   "basePersistDirectory": "\/tmp\/realtime\/basePersist"
  }
}
{
 "queryType": "groupBy",
 "dataSource": "pci",
 "granularity": "none",
 "limitSpec": { "type": "default", "limit": 500000000, "columns": [] },
 "dimensions": ["account_number","tcode","timezone"],
 "intervals" : [ "2012-01-01T00:00:00.000/2014-10-01T00:00:00.000" ],
  "aggregations":[
          {"type":"longSum","name":"count","fieldName":"count"},
          {"type":"longSum","name":"amount","fieldName":"amount"}
    ]
}
...
--
You received this message because you are subscribed to the Google Groups "Druid Development" group.
To unsubscribe from this group and stop receiving emails from it, send an email to druid-developm...@googlegroups.com.
To post to this group, send email to druid-de...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/druid-development/87564bbd-72fa-4d8c-aad2-648fd3e62d16%40googlegroups.com.
[sudhrao@dtee1-druid-poc-001-6055 ~]$ cat /opt/druid/run_historical.shÂ
#!/usr/bin/env bash
echo "This will run a historical of Druid"
set +u
shopt -s xpg_echo
shopt -s expand_aliases
#Â start process
JAVA_ARGS="-Xmx1g -Duser.timezone=UTC -Dfile.encoding=UTF-8"
#For the kit
DRUID_CP=/opt/druid/lib/*:/opt/druid/config/historical
echo "Running command:"
(set -x; java -XX:MaxDirectMemorySize=8g ${JAVA_ARGS} -classpath ${DRUID_CP} -Ddruid.emitter=logging -Ddruid.emitter.logging.logLevel=error io.druid.cli.Main server historical)
[sudhrao@dtee1-druid-poc-001-6055 ~]$Â
similarly other servers, with mx set to 1g-4g
...
{
  "queryType" : "timeBoundary",
  "dataSource": "pci"
}
Result :
[ {
 "timestamp" : "2012-01-01T00:12:37.000Z",
 "result" : {
  "minTime" : "2012-01-01T00:12:37.000Z",
  "maxTime" : "2013-09-30T21:58:55.000Z"
 }
} ]
...
...
...