Filip Dziuba
未讀,2021年9月23日 上午11:25:392021/9/23登入以回覆作者
登入以轉寄訊息
你的權限不足,無法在這個群組刪除訊息
登入以檢舉訊息
該群組的電子郵件地址為匿名,或你需要檢視成員電子郵件地址的權限才能查看原始貼文
收件者:Druid User
Hi,
I am trying to reindex data and change query granularity. I am reading from Druid data source A and writing to data source B.
Ingestion is separated into 20 subtasks with 10 running at the same time. One subtask usually fails but on retry it succeeded so the whole index_parallel task is successful.
I can see that segment and partitions are created, Druid UI see the number of rows, but query return no data. There are no errors on any node.
I am quite stuck, the spec work on smaller data set.
The only worrying thing I can see is that when one subtask fails the partition it was supposed to create is never created. The retry will create new partition. So I can have set of 20 partitions 0-20 with 12th partition missing, so count is OK, but they are not in sequence.
Is this a bug or should druid handle partitions with missing numbers?