[SERVER-24035] balancer does not respect active window Created: 03/May/16 Updated: 04/May/16 Resolved: 04/May/16 |
|
| Status: | Closed |
| Project: | Core Server |
| Component/s: | Sharding |
| Affects Version/s: | 3.0.5 |
| Fix Version/s: | None |
| Type: | Bug | Priority: | Major - P3 |
| Reporter: | Dai Shi | Assignee: | Unassigned |
| Resolution: | Duplicate | Votes: | 0 |
| Labels: | None | ||
| Remaining Estimate: | Not Specified | ||
| Time Spent: | Not Specified | ||
| Original Estimate: | Not Specified | ||
| Issue Links: |
|
||||||||
| Operating System: | ALL | ||||||||
| Participants: | |||||||||
| Description |
|
We just turned the balancer back on for a cluster where it has been off for many months. Typically, we run the balancer 24/7, but in this case we can only run it for a few hours a day. I configured the balancer active window to be between 23:00 and 6:00, per the example given here: https://docs.mongodb.org/v3.0/tutorial/manage-sharded-cluster-balancer/#schedule-the-balancing-window However, I noticed that the balancer was running even outside of these hours, and actually caused severe issues to our site this morning. I've reproduced the steps here:
As you can see, the time when I ran those commands was just after 21:00, which should not be inside the active window. However, after turning the balancer on, it immediately started migrating chunks. Is there something I'm missing? |
| Comments |
| Comment by Dai Shi [ 04/May/16 ] | ||
|
OK, thanks for looking into this. We will patch upgrade before turning the balancer back on. | ||
| Comment by Ramon Fernandez Marina [ 04/May/16 ] | ||
|
dai@foursquare.com, this issue was reported earlier in Please consider upgrading to the latest 3.0 release (3.0.11 at the time of this writing) at your earliest convenience. Note also that 3.2 is not affected by this issue (3.2.6 is the latest release in that branch). Thanks, | ||
| Comment by Ramon Fernandez Marina [ 04/May/16 ] | ||
|
dai@foursquare.com, I'm able to reproduce what I think is the same behavior you describe in a test cluster. I've also seem to have found a workaround though – after you enable the balancer, disable it again and re-enable it:
We're investigating and we'll post updates on this ticket when we have them. Thanks, |