You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Segments becoming frequently unavailable when replica = 1 for large datasource
Affected Version
26.0.0
Description
We have noticed that one of the data source which has 3 TB of data having 30K segments is having frequently unavailable segments. From the finding it looks to me, it is a coordinator balancing issue, where coordinator load a segment to new historical and after loading on new one it ends up dropping from both place.
Not sure if i have any steps to reproduce it. As this happens when coordinator does the re-balancing.
Finding
This is what we have observed in the logs for a specific segment. Let me know if complete logs needed i will try to get it.
coordinator asks a new historical to load the segment.
next it ask the same new historical to drop the segment which it just loaded because it sees that replica =2.
next it ask the older historical to drop the data, as i am assuming some callback went in saying that new node has loaded the segment so it should also drop.
The text was updated successfully, but these errors were encountered:
This issue has been marked as stale due to 280 days of inactivity.
It will be closed in 4 weeks if no further activity occurs. If this issue is still
relevant, please simply write any comment. Even if closed, you can still revive the
issue at any time or discuss it on the dev@druid.apache.org list.
Thank you for your contributions.
This issue has been closed due to lack of activity. If you think that
is incorrect, or the issue requires additional review, you can revive the issue at
any time.
Segments becoming frequently unavailable when replica = 1 for large datasource
Affected Version
26.0.0
Description
We have noticed that one of the data source which has 3 TB of data having 30K segments is having frequently unavailable segments. From the finding it looks to me, it is a coordinator balancing issue, where coordinator load a segment to new historical and after loading on new one it ends up dropping from both place.
This is what we have observed in the logs for a specific segment. Let me know if complete logs needed i will try to get it.
The text was updated successfully, but these errors were encountered: