Storage node failure state seen after unlock
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
StarlingX |
Fix Released
|
Medium
|
Daniel Badea |
Bug Description
Brief Description
-----------------
After unlocking a storage node, it was observed to go into failed state. It was rebooted and then came up as available.
Severity
--------
Major
Steps to Reproduce
------------------
1. Ceph was added as a backend via system storage-backend-add
2. Storage nodes were added to the system
3. After the storage nodes were in locked/online state, the nodes were unlocked
4. On unlock, both storage nodes were observed to go into failed state.
5. They were rebooted and after reboot, they became available.
6. Looking at the logs, I see a number of failures in ceph-osd-prepare as follows:
2018-10-
2018-10-
It should be noted that a 'sudo wipedisk' was run prior to the storage nodes being provisioned.
Expected Behavior
------------------
The storage nodes do not go into available state.
Actual Behavior
----------------
Storage nodes report failed state, reboot and then become available.
Reproducibility
---------------
Tried once so far but seen on both storage nodes.
System Configuration
-------
Storage
Branch/Pull Time/Commit
-------
master as of 2018-10-24_21-18-00
Timestamp/Logs
--------------
See above
tags: |
added: stx.2019.05 removed: stx.2019.03 |
tags: |
added: stx.2.0 removed: stx.2019.05 |
Duplicate of https:/ /bugs.launchpad .net/starlingx/ +bug/1800889