The duplicated LP: 1827119 has fix released with patch: https://review.opendev.org/#/c/677424/ which merged in Sep 23th, 2019.
I have tried with 20191120T023000Z to verify the issue, but cannot reproduce, since set ceph_mon_gib=23 cannot return sucess because the available space in compute-0 have only 1 GB growth.
@Wendy, can you re-check the issue with latest image?
Following is the print information from my evaluation:
[sysadmin@controller-0 ~(keystone_admin)]$ system ceph-mon-list
+--------------------------------------+--------------+--------------+------------+------+
| uuid | ceph_mon_gib | hostname | state | task |
+--------------------------------------+--------------+--------------+------------+------+
| 2e78c9e8-e6b8-4f79-b498-92de29538c24 | 20 | controller-0 | configured | None |
| 372658d0-0a44-465a-9d8f-12d6dba6b38e | 20 | controller-1 | configured | None |
| 53a8dce1-e830-4514-97f2-652a9ec5ab9f | 20 | compute-0 | configured | None |
+--------------------------------------+--------------+--------------+---------c
$ system ceph-mon-modify controller-0 ceph_mon_gib=40
Node: compute-0 Total target growth size 20 GiB for database (doubled for upgrades), glance, scratch, backup, extension and ceph-mon exceeds growth limit of 1 GiB.
$ system ceph-mon-modify controller-0 ceph_mon_gib=23
Node: compute-0 Total target growth size 3 GiB for database (doubled for upgrades), glance, scratch, backup, extension and ceph-mon exceeds growth limit of 1 GiB.
The duplicated LP: 1827119 has fix released with patch: https:/ /review. opendev. org/#/c/ 677424/ which merged in Sep 23th, 2019.
I have tried with 20191120T023000Z to verify the issue, but cannot reproduce, since set ceph_mon_gib=23 cannot return sucess because the available space in compute-0 have only 1 GB growth.
@Wendy, can you re-check the issue with latest image?
Following is the print information from my evaluation: controller- 0 ~(keystone_admin)]$ system ceph-mon-list ------- ------- ------- ------- ----+-- ------- -----+- ------- ------+ ------- -----+- -----+ ------- ------- ------- ------- ----+-- ------- -----+- ------- ------+ ------- -----+- -----+ e6b8-4f79- b498-92de29538c 24 | 20 | controller-0 | configured | None | 0a44-465a- 9d8f-12d6dba6b3 8e | 20 | controller-1 | configured | None | e830-4514- 97f2-652a9ec5ab 9f | 20 | compute-0 | configured | None | ------- ------- ------- ------- ----+-- ------- -----+- ------- ------+ ------- --c
[sysadmin@
+------
| uuid | ceph_mon_gib | hostname | state | task |
+------
| 2e78c9e8-
| 372658d0-
| 53a8dce1-
+------
$ system ceph-mon-modify controller-0 ceph_mon_gib=40
Node: compute-0 Total target growth size 20 GiB for database (doubled for upgrades), glance, scratch, backup, extension and ceph-mon exceeds growth limit of 1 GiB.
$ system ceph-mon-modify controller-0 ceph_mon_gib=23
Node: compute-0 Total target growth size 3 GiB for database (doubled for upgrades), glance, scratch, backup, extension and ceph-mon exceeds growth limit of 1 GiB.
$ system ceph-mon-modify controller-0 ceph_mon_gib=21 ------- ------- ------- ------- ----+-- ------- -----+- ------- ------+ ------- -----+- -----+ ------- ------- ------- ------- ----+-- ------- -----+- ------- ------+ ------- -----+- -----+ e6b8-4f79- b498-92de29538c 24 | 21 | controller-0 | configured | None | 0a44-465a- 9d8f-12d6dba6b3 8e | 21 | controller-1 | configured | None | e830-4514- 97f2-652a9ec5ab 9f | 21 | compute-0 | configured | None | ------- ------- ------- ------- ----+-- ------- -----+- ------- ------+ ------- -----+- -----+
+------
| uuid | ceph_mon_gib | hostname | state | task |
+------
| 2e78c9e8-
| 372658d0-
| 53a8dce1-
+------
NOTE: ceph_mon_gib for both controllers are changed.
System configuration has changed.
please follow the administrator guide to complete configuring system.
# After compute-0 lock and unlock, the system host-list shows normally.
[sysadmin@ controller- 0 ~(keystone_admin)]$ system host-list ------- ------+ ------- ------+ ------- ------- --+---- ------- --+---- ------- ---+ ------- ------+ ------- ------+ ------- ------- --+---- ------- --+---- ------- ---+ ------- ------+ ------- ------+ ------- ------- --+---- ------- --+---- ------- ---+
+----+-
| id | hostname | personality | administrative | operational | availability |
+----+-
| 1 | controller-0 | controller | unlocked | enabled | available |
| 2 | controller-1 | controller | unlocked | enabled | available |
| 3 | compute-0 | worker | unlocked | enabled | available |
| 4 | compute-1 | worker | unlocked | enabled | available |
+----+-
# ssh to compute-0 and check: -vg-scratch- -lv 253:1 0 3.9G 0 lvm /scratch -vg-kubelet- -lv 253:3 0 10G 0 lvm /var/lib/kubelet -vg-ceph- -mon--lv 253:4 0 21G 0 lvm /var/lib/ceph/mon -vg-docker- -lv 253:5 0 30G 0 lvm /var/lib/docker -local- instances_ lv 253:0 0 10G 0 lvm /var/lib/ nova/instances
compute-0:~$ lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 200G 0 disk
├─sda1 8:1 0 1M 0 part
├─sda2 8:2 0 500M 0 part /boot
├─sda3 8:3 0 69G 0 part
│ ├─cgts-
│ ├─cgts--vg-log--lv 253:2 0 3.9G 0 lvm /var/log
│ ├─cgts-
│ ├─cgts-
│ └─cgts-
├─sda4 8:4 0 19.5G 0 part /
├─sda5 8:5 0 10G 0 part
│ └─nova-
├─sda6 8:6 0 30G 0 part
├─sda7 8:7 0 20G 0 part
└─sda8 8:8 0 20G 0 part
sdb 8:16 0 30G 0 disk