The image used is: 20211201T041648Z [sysadmin@controller-0 ~(keystone_admin)]$ fm alarm-list +----------+---------------------------------------------------------------------------------------+--------------------------------------+----------+-------------------+ | Alarm ID | Reason Text | Entity ID | Severity | Time Stamp | +----------+---------------------------------------------------------------------------------------+--------------------------------------+----------+-------------------+ | 800.010 | Potential data loss. No available OSDs in storage replication group group-0: no OSDs | cluster=07c0ce0c-9b00-4de0-a8ee- | critical | 2021-12-01T08:36: | | | | 13da4debf43c.peergroup=group-0 | | 39.009657 | | | | | | | | 800.001 | Storage Alarm Condition: HEALTH_WARN. Please check 'ceph -s' for more details. | cluster=07c0ce0c-9b00-4de0-a8ee- | warning | 2021-12-01T08:14: | | | | 13da4debf43c | | 14.949030 | | | | | | | | 250.001 | compute-1 Configuration is out-of-date. | host=compute-1 | major | 2021-12-01T08:07: | | | | | | 16.063014 | | | | | | | | 250.001 | compute-0 Configuration is out-of-date. | host=compute-0 | major | 2021-12-01T08:07: | | | | | | 13.642871 | | | | | | | | 200.001 | compute-1 was administratively locked to take it out-of-service. | host=compute-1 | warning | 2021-12-01T07:56: | | | | | | 59.093516 | | | | | | | | 200.001 | compute-0 was administratively locked to take it out-of-service. | host=compute-0 | warning | 2021-12-01T07:56: | | | | | | 53.223718 | | | | | | | +----------+---------------------------------------------------------------------------------------+--------------------------------------+----------+-------------------+ [sysadmin@controller-0 ~(keystone_admin)]$ system host-list +----+--------------+-------------+----------------+-------------+--------------+ | id | hostname | personality | administrative | operational | availability | +----+--------------+-------------+----------------+-------------+--------------+ | 1 | controller-0 | controller | unlocked | enabled | available | | 2 | controller-1 | controller | unlocked | enabled | available | | 3 | compute-0 | worker | locked | disabled | online | | 4 | compute-1 | worker | locked | disabled | online | +----+--------------+-------------+----------------+-------------+--------------+ [sysadmin@controller-0 ~(keystone_admin)]$ system host-unlock 3 A host with worker functionality requires a nova-local volume group prior to being enabled.The nova-local volume group does not contain any physical volumes in the adding or provisioned state. [sysadmin@controller-0 ~(keystone_admin)]$ collect all [sudo] password for sysadmin: collecting data from 4 host(s) Error: cannot collect from compute-0 (reason:33:unreachable) Error: cannot collect from compute-1 (reason:33:unreachable) monitoring host collect ; please standby .. collected controller-1_20211201.085016 ... done (00:00:40 16M 1%) collected controller-0_20211201.085016 ... done (00:01:41 19M 1%) collected from 2 hosts creating all-nodes tarball /scratch/ALL_NODES_20211201.085016.tar ... done (00:01:42 35M 1%) [sysadmin@controller-0 ~(keystone_admin)]$ [sysadmin@controller-0 ~(keystone_admin)]$ ceph -s cluster: id: 07c0ce0c-9b00-4de0-a8ee-13da4debf43c health: HEALTH_WARN OSD count 0 < osd_pool_default_size 2
services: mon: 2 daemons, quorum controller-0,controller-1 (age 22m) mgr: controller-0(active, since 44m), standbys: controller-1 mds: 2 up:standby osd: 0 osds: 0 up, 0 in
data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs:
The image used is: 20211201T041648Z controller- 0 ~(keystone_admin)]$ fm alarm-list ----+-- ------- ------- ------- ------- ------- ------- ------- ------- ------- ------- ------- ------- -+----- ------- ------- ------- ------- -----+- ------- --+---- ------- ------- -+ ----+-- ------- ------- ------- ------- ------- ------- ------- ------- ------- ------- ------- ------- -+----- ------- ------- ------- ------- -----+- ------- --+---- ------- ------- -+ 07c0ce0c- 9b00-4de0- a8ee- | critical | 2021-12-01T08:36: | peergroup= group-0 | | 39.009657 | 07c0ce0c- 9b00-4de0- a8ee- | warning | 2021-12-01T08:14: | ----+-- ------- ------- ------- ------- ------- ------- ------- ------- ------- ------- ------- ------- -+----- ------- ------- ------- ------- -----+- ------- --+---- ------- ------- -+ controller- 0 ~(keystone_admin)]$ system host-list ------- ------+ ------- ------+ ------- ------- --+---- ------- --+---- ------- ---+ ------- ------+ ------- ------+ ------- ------- --+---- ------- --+---- ------- ---+ ------- ------+ ------- ------+ ------- ------- --+---- ------- --+---- ------- ---+ controller- 0 ~(keystone_admin)]$ system host-unlock 3 controller- 0 ~(keystone_admin)]$ collect all 33:unreachable) 33:unreachable) 1_20211201. 085016 ... done (00:00:40 16M 1%) 0_20211201. 085016 ... done (00:01:41 19M 1%) ALL_NODES_ 20211201. 085016. tar ... done (00:01:42 35M 1%) controller- 0 ~(keystone_admin)]$ controller- 0 ~(keystone_admin)]$ ceph -s 9b00-4de0- a8ee-13da4debf4 3c default_ size 2
[sysadmin@
+------
| Alarm ID | Reason Text | Entity ID | Severity | Time Stamp |
+------
| 800.010 | Potential data loss. No available OSDs in storage replication group group-0: no OSDs | cluster=
| | | 13da4debf43c.
| | | | | |
| 800.001 | Storage Alarm Condition: HEALTH_WARN. Please check 'ceph -s' for more details. | cluster=
| | | 13da4debf43c | | 14.949030 |
| | | | | |
| 250.001 | compute-1 Configuration is out-of-date. | host=compute-1 | major | 2021-12-01T08:07: |
| | | | | 16.063014 |
| | | | | |
| 250.001 | compute-0 Configuration is out-of-date. | host=compute-0 | major | 2021-12-01T08:07: |
| | | | | 13.642871 |
| | | | | |
| 200.001 | compute-1 was administratively locked to take it out-of-service. | host=compute-1 | warning | 2021-12-01T07:56: |
| | | | | 59.093516 |
| | | | | |
| 200.001 | compute-0 was administratively locked to take it out-of-service. | host=compute-0 | warning | 2021-12-01T07:56: |
| | | | | 53.223718 |
| | | | | |
+------
[sysadmin@
+----+-
| id | hostname | personality | administrative | operational | availability |
+----+-
| 1 | controller-0 | controller | unlocked | enabled | available |
| 2 | controller-1 | controller | unlocked | enabled | available |
| 3 | compute-0 | worker | locked | disabled | online |
| 4 | compute-1 | worker | locked | disabled | online |
+----+-
[sysadmin@
A host with worker functionality requires a nova-local volume group prior to being enabled.The nova-local volume group does not contain any physical volumes in the adding or provisioned state.
[sysadmin@
[sudo] password for sysadmin:
collecting data from 4 host(s)
Error: cannot collect from compute-0 (reason:
Error: cannot collect from compute-1 (reason:
monitoring host collect ; please standby ..
collected controller-
collected controller-
collected from 2 hosts
creating all-nodes tarball /scratch/
[sysadmin@
[sysadmin@
cluster:
id: 07c0ce0c-
health: HEALTH_WARN
OSD count 0 < osd_pool_
services: 0,controller- 1 (age 22m) 0(active, since 44m), standbys: controller-1
mon: 2 daemons, quorum controller-
mgr: controller-
mds: 2 up:standby
osd: 0 osds: 0 up, 0 in
data:
pools: 0 pools, 0 pgs
objects: 0 objects, 0 B
usage: 0 B used, 0 B / 0 B avail
pgs: