some storage nodes staying unlocked and cannot be locked/force-locked after controller-0/system restored
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
StarlingX |
Fix Released
|
Medium
|
Wei Zhou |
Bug Description
Brief Description
-----------------
After powered on and ran 'config_controller --restore-system <bk_system...>' on the controller-0, some of the storage nodes were in 'unlocked | enabled | degraded' states.
Attempt to lock them failed with error:
[wrsroot@
Cannot lock a storage node when ceph pool usage is undetermined.
[wrsroot@
1
Force-lock also failed:
[wrsroot@
Cannot lock a storage node when ceph pool usage is undetermined.
Severity
--------
Major
Steps to Reproduce
------------------
1 backup the system and save the backup files
2 install the controller-0 with the same load (powered off all other nodes first)
3 run 'sudo config_controller --restore-system <system_
4 source /etc/nova/openrc
5 lock any nodes (exception controller-0) if they are unlocked, use 'force-lock' if 'lock' is not working
Expected Behavior
------------------
All nodes should be able to be locked, or can be locked, or can be force-locked.
Actual Behavior
----------------
For the unlocked storage nodes, 'system host-lock' failed with error message, e.g. for storage-1:
[wrsroot@
Cannot lock a storage node when ceph pool usage is undetermined.
Force-lock also failed:
[wrsroot@
Cannot lock a storage node when ceph pool usage is undetermined.
Reproducibility
---------------
Reproducible on pv0
System Configuration
-------
Dedicated storage 2 + 6 + 4
Branch/Pull Time/Commit
-------
StarlingX_18.10 as of 2018-10-12_01-52-00
Timestamp/Logs
--------------
20181016 16:56:05
tags: |
added: stx.2019.05 removed: stx.2019.03 |
tags: |
added: stx.2.0 removed: stx.2019.05 |
Targeting stx.2019.03 as this is specific to a 6-storage node config which is not very common. It was confirmed that this is not an issue for 2-storage node configs.