2015-03-20T01:00:00.000Z_2015-03-20T02:00:00.000Z
2015-03-23T02:00:00.000Z_2015-03-23T03:00:00.000Z2015-03-23T01:00:00.000Z_2015-03-23T02:00:00.000Z
2015-03-23T03:00:00.000Z_2015-03-23T04:00:00.000Z
# Extensionsdruid.extensions.coordinates=["io.druid.extensions:druid-kafka-eight","io.druid.extensions:mysql-metadata-storage"]druid.extensions.remoteRepositories=["http://repo1.maven.org/maven2/", "https://metamx.artifactoryonline.com/metamx/pub-libs-releases-local"]
# Zookeeperdruid.zk.service.host=10.200.6.60:2181,10.200.6.70:2181druid.zk.paths.base=/druid
# Metadata Storage (mysql)druid.metadata.storage.type=mysqldruid.metadata.storage.connector.connectURI=jdbc\:mysql\://10.200.6.5\:3306/druiddruid.metadata.storage.connector.user=druiddruid.metadata.storage.connector.password=druid
# Deep storage (local filesystem for examples - don't use this in production)druid.storage.type=localdruid.storage.storage.storageDirectory=/nas/druid/dataStorage
# Query Cache (we use a simple 10mb heap-based local cache on the broker)druid.cache.type=localdruid.cache.sizeInBytes=10000000
# Indexing service discoverydruid.selectors.indexing.serviceName=overlord
# Monitoring (disabled for examples)# druid.monitoring.emissionPeriod = PT5m# druid.monitoring.monitors=["com.metamx.metrics.SysMonitor","com.metamx.metrics.JvmMonitor"]
# Metrics logging (disabled for examples)druid.emitter=loggingdruid.emitter.logging.loggerClass=LoggingEmitterdruid.emitter.logging.logLevel=debug
[{ "dataSchema" : { "dataSource" : "buck_bidding", "parser" : { "type" : "string", "parseSpec" : { "format" : "json", "timestampSpec" : { "column" : "insert_datetime", "format" : "yyyy-MM-dd HH:mm:ss" }, "dimensionsSpec" : { "dimensions": ["bid","win","click","convert","exchange_name","creative_id","device_country","device_make","device_model","device_os","device_osv","device_type","carrierid","bid_price","win_price","company_price","advertiser_price","pub_id","pub_name","place_id","place_name","domain","categories","exchange_id","campaign_id","offer_id","advertiser_id","device_country_id","device_make_id","device_model_id","device_os_id","device_osv_id","device_type_id","carrierid_id","banner_width","banner_heigth","supply_type","session_bid_id"], "dimensionExclusions" : [], "spatialDimensions" : [] } } }, "metricsSpec" : [{ "type" : "count", "name" : "count" },{ "type" : "doubleSum", "name" : "bid_spent", "fieldName": "bid_price" },{ "type": "longSum", "name" : "bid_response", "fieldName" : "bid" },{ "type": "longSum", "name" : "win", "fieldName" : "win" } ], "granularitySpec" : { "type" : "uniform", "segmentGranularity" : "hour", "queryGranularity" : "NONE" } }, "ioConfig" : { "type" : "realtime", "firehose": { "type": "kafka-0.8", "consumerProps": { "zookeeper.connect": "10.200.6.60:2181,10.200.6.70:2181", "zookeeper.connection.timeout.ms" : "15000", "zookeeper.session.timeout.ms" : "15000", "zookeeper.sync.time.ms" : "5000", "group.id": "druid", "fetch.message.max.bytes" : "1048586", "auto.offset.reset": "largest", "auto.commit.enable": "true" }, "feed": "buck_bidding" }, "plumber": { "type": "realtime" } }, "tuningConfig": { "type" : "realtime", "maxRowsInMemory": 500000, "intermediatePersistPeriod": "PT10m", "windowPeriod": "PT2h", "basePersistDirectory": "\/nas\/druid\/dataStorage", "rejectionPolicy": { "type": "messageTime" } } }]
druid.host=10.200.6.64druid.service=realtimedruid.port=8083
druid.processing.buffer.sizeBytes=100000000druid.processing.numThreads=3
druid.server.tier=tats_tier
druid.monitoring.monitors=["io.druid.segment.realtime.RealtimeMetricsMonitor"]
druid.realtime.specFile=/usr/local/druid/config/realtime/bucksense_new_realtime.spec
druid.service=overlorddruid.host=10.200.6.60druid.port=8087
# Task Log Module (Overlord and MiddleManager node)druid.indexer.logs.type=filedruid.indexer.logs.directory=/usr/local/druid/logs/
druid.indexer.queue.startDelay=PT1M
druid.indexer.runner.type=remotedruid.indexer.storage.type=metadata
druid.service=middleManagerdruid.host=10.200.6.62druid.port=8091
# Task Log Module (Overlord and MiddleManager node)druid.indexer.logs.type=filedruid.indexer.logs.directory=/usr/local/druid/logs/
# Middle Manager Configurationdruid.worker.ip=10.200.6.62druid.worker.capacity=1
# Peon Configurationdruid.indexer.runner.javaOpts="-server -Xmx3g -XX:+UseGxi2GC -XX:MaxGCPauseMillis=100 -XX:+PrintGCDetails -XX:+PrintGCTimeStamps"druid.indexer.task.chathandler.type=announcedruid.indexer.task.baseTaskDir=/nas/druid/task
druid.indexer.fork.property.druid.computation.buffer.size=536870912druid.indexer.fork.property.druid.processing.numThreads=3druid.indexer.fork.property.druid.request.logging.type=filedruid.indexer.fork.property.druid.request.logging.dir=/usr/local/druid/logsdruid.indexer.fork.property.druid.segmentCache.locations=[{"path": "/nas/druid/indexCache", "maxSize": 0}]druid.indexer.fork.property.druid.server.http.numThreads=50
druid.host=10.200.6.62druid.service=brokerdruid.port=8080
# Druid Processing Module (Historical, Realtime, and Broker nodes)druid.processing.numThreads=7druid.processing.buffer.sizeBytes=100000000
# Queryable Module (Historical, Realtime, and Broker nodes)druid.request.logging.type=filedruid.request.logging.dir=/usr/local/druid/logs
druid.host=10.200.6.70druid.port=8081druid.service=historical
# Druid Processing Module (Historical, Realtime, and Broker nodes)druid.processing.numThreads=2druid.processing.buffer.sizeBytes=536870912
druid.historical.cache.useCache=truedruid.historical.cache.populateCache=true
druid.server.maxSize=10000000000druid.server.tier=tats_tierdruid.server.priority=50druid.segmentCache.locations=[{"path": "/nas/druid/indexCache", "maxSize": 10000000000}]
Hi Fangjin,
I'm doing queries to the Broker module.
Into Coordinator console I can see all segments belongs to Realtime server but the one I've submitted on Friday are anymore there.
I've also noticed that segments are not stored on Mysql druid_segments table, can it be an issue?
I didn't find any error on Realtime logs, did you have any specific suggestion I can search for?
I've set messageTime because I'm currently simulating my traffic so timestamp doesn't match with server one.
Did you think it can generate some kind of issue on data movement from Realtime to Historical?
Just to be sure, setting the storageDirectory to the same path for both Realtime and Historical, is it the right thing?
Thanks
Maurizio
sent by Nexus
--
You received this message because you are subscribed to a topic in the Google Groups "Druid Development" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/druid-development/6VD98RDgsZI/unsubscribe.
To unsubscribe from this group and all its topics, send an email to druid-developm...@googlegroups.com.
To post to this group, send email to druid-de...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/druid-development/d94fcd9c-c5e7-4466-8cb8-dc2fd1964980%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
Hi Fangjin,
I'm doing queries to the Broker module.
Into Coordinator console I can see all segments belongs to Realtime server but the one I've submitted on Friday are anymore there.
I've also noticed that segments are not stored on Mysql druid_segments table, can it be an issue?
I didn't find any error on Realtime logs, did you have any specific suggestion I can search for?
I've set messageTime because I'm currently simulating my traffic so timestamp doesn't match with server one.
Did you think it can generate some kind of issue on data movement from Realtime to Historical?
Just to be sure, setting the storageDirectory to the same path for both Realtime and Historical, is it the right thing?
To unsubscribe from this group and all its topics, send an email to druid-development+unsubscribe@googlegroups.com.
To post to this group, send email to druid-development@googlegroups.com.
[root@druid06 buck_bidding]# ls2015-03-23T01:00:00.000Z_2015-03-23T02:00:00.000Z 2015-03-24T11:00:00.000Z_2015-03-24T12:00:00.000Z[root@druid06 buck_bidding]# cd 2015-03-24T11\:00\:00.000Z_2015-03-24T12\:00\:00.000Z/[root@druid06 2015-03-24T11:00:00.000Z_2015-03-24T12:00:00.000Z]# ls0 1 2 3 4 5
Exception in thread "plumber_persist_4" com.metamx.common.ISE: Cannot merge columns of type[STRING] and [LONG] at io.druid.segment.column.ColumnCapabilitiesImpl.merge(ColumnCapabilitiesImpl.java:124) at io.druid.segment.IndexMerger.makeIndexFiles(IndexMerger.java:423) at io.druid.segment.IndexMerger.merge(IndexMerger.java:319) at io.druid.segment.IndexMerger.persist(IndexMerger.java:181) at io.druid.segment.IndexMerger.persist(IndexMerger.java:151) at io.druid.segment.IndexMerger.persist(IndexMerger.java:134) at io.druid.segment.realtime.plumber.RealtimePlumber.persistHydrant(RealtimePlumber.java:830) at io.druid.segment.realtime.plumber.RealtimePlumber$3.doRun(RealtimePlumber.java:354) at io.druid.common.guava.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:40) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)
Exception in thread "plumber_persist_5" com.metamx.common.ISE: Cannot merge columns of type[STRING] and [LONG] at io.druid.segment.column.ColumnCapabilitiesImpl.merge(ColumnCapabilitiesImpl.java:124) at io.druid.segment.IndexMerger.makeIndexFiles(IndexMerger.java:423) at io.druid.segment.IndexMerger.merge(IndexMerger.java:319) at io.druid.segment.IndexMerger.persist(IndexMerger.java:181) at io.druid.segment.IndexMerger.persist(IndexMerger.java:151) at io.druid.segment.IndexMerger.persist(IndexMerger.java:134) at io.druid.segment.realtime.plumber.RealtimePlumber.persistHydrant(RealtimePlumber.java:830) at io.druid.segment.realtime.plumber.RealtimePlumber$3.doRun(RealtimePlumber.java:354) at io.druid.common.guava.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:40) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)
Here again on metadata (mysql) I didn't find segment row info even if it's shown into Coordinator console.I think I need to work on my incoming data checking what's wrong.Can you please clarify when"basePersistDirectory" and "druid.storage.storageDirectory" are involved, seems that Reatime Node has only used basePersistDirectory
In your previous thread you mention "5. A historical locally downloads the segment from deep storage and serves the segment." where Historical will store it, how can I setup this path?
{ "type": "longSum", "name" : "bid_response", "fieldName" : "bid" },{ "type": "longSum", "name" : "win", "fieldName" : "win"
},{ "type": "longSum", "name" : "clicks", "fieldName" : "click" },{ "type": "longSum", "name" : "conversion", "fieldName" : "convert" }