Thanks a lot David, will try and let you know the results :)
--
You received this message because you are subscribed to a topic in the Google Groups "Druid User" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/druid-user/25MQIyqZqUI/unsubscribe.
To unsubscribe from this group and all its topics, send an email to druid-user+unsubscribe@googlegroups.com.
To post to this group, send email to druid...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/druid-user/b2652179-ae53-412a-8965-260abf35d2cc%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
Hi David,
One more help was required – wanted to understand the exact use or significance of segment-cache. I am quite confused by reading multiple documents and threads on segment-cache.
Segment-cache are those segments which are not loaded yet into deep storage ? OR
Is it something which was extracted from deep storage to serve query ? OR
Druid keep all the segments by default on local storage between on historical nodes ?
Also what does Pending Segments signify ?
Thanks in advance.
Regards,
Arpan Khagram
Mobile: +91 8308993200
To unsubscribe from this group and all its topics, send an email to druid-user+...@googlegroups.com.
Thanks David for the quick response J
With respect to same question, how multiple historical nodes are supposed to work with respect to segment-cache. If I have configured 2 historical, both of them are supposed to have same dataset in segment-cache ? How having multiple historical nodes in my cluster helps me ?
Mobile: +91 8308993200
Hi David,CommunityThanks for the reply :) Few more things -1. 400-500 MB is recommended size for the complete segment or the partition ? What if I have 2-3 GB size of segment but partition is only 80-90 MB. Is this good to have or should change anything ?
2. What exactly segment loading means ? I wanted to understand the complete flow of the data for KAFKA indexing Service.
I can see segments getting loaded first on Deep Storage and then it comes to segment-cache - is this correct understanding ?
3. I have configured 2 historical nodes and I can see both are having same segment-cache. I understood from your below statement and druid docs that it should load balanceIn the general case, the segment caches of your historical nodes will not contain the same segments.
Server maxSize sets the maximum cumulative segment size (in bytes) that a node can hold. Changing this parameter will affect performance by controlling the memory/disk ratio on a node. Setting this parameter to a value greater than the total memory capacity on a node and may cause disk paging to occur. This paging time introduces a query latency delay.
Hi David,
Wanted to Check if we can force historic nodes to specific segments mapping (date wise) - such that we can have older segments go on nodes which have a higher druid.server.maxSize.
Example - Is there a way i can keep current(latest) 5 days of data on few historical nodes and data older than 5 days on another historical nodes.
Reason - Most of my queries will be for recent 4/5 days of data, i can manage to get servers(VMs) with large RAM (memory) so that queries are faster. It's OK for me if it takes little bit of time while querying older data ( i can keep older data in separate historical nodes with low RAM as well). I have requirement to keep 28 days of data and it is becoming difficult to get so many large servers which can accommodate all days of data.
Also is there any way i can force some amount of data to be present permanently (says last 1/2/3 days of data) in memory for query.