-
Notifications
You must be signed in to change notification settings - Fork 31
How to simulate most monitor failure scenarios? #323
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
@haoxiaoci it should be temporary state |
it has been probing for more than 10min.... |
In that case try removing mons from cluster and then check https://docs.ceph.com/en/latest/rados/operations/add-or-rm-mons/#removing-a-monitor-manual |
after scale mon-c, mon-d to 0, ceph command execute got stuck, so i can't rm mon c or d manually. |
@haoxiaoci you can try to scale mon to 1 in cluster.yaml config file. or you can first try to scale up the mon pods and try again. |
@subhamkrai I wonder if the tool should also allow resetting the quorum in case the state is |
not all |
thanks for the feedback, let's continue the discussion on #326. |
Uh oh!
There was an error while loading. Please reload this page.
How to simulate most of the monitor failure scenarios?
in my test environment, i have 3 health mons, and rook-operator works fine, ceph cluster is and found mon-a is leader
in order to simulate majority mons offline, i scale down mon-d and mon-c
then check mon-a state, found it's in
probling
statethen i can't use restore-quorum to restore c, d from good mon a.
The text was updated successfully, but these errors were encountered: