VM build fails after Train-Ussuri upgrade
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
kolla-ansible |
Fix Released
|
Undecided
|
Unassigned |
Bug Description
What happened:
I upgraded my Train test cluster to Ussuri following these instructions:
https:/
The upgrade completed successfully with no failures, and the existing VMs are fine, but new VM build fails with rados.Rados.
http://
I'm running external ceph so I looked at this document:
OpenStack Docs: External Ceph<https:/
It says that I need the following in /etc/kolla/
auth_cluster_
auth_service_
auth_client_
I didn't have that, so I added it and then redeployed, but still can't build VMs. I tried adding the same to all copies of ceph.conf and redeployed again, but that didn't help. I see some cryptic talk about ceph in the release notes but it's not obvious what I'm being asked to change:
I read the bug that it refers to:
https:/
But I already have "backend_
Also I read these sections but I don't see anything obvious here that needs to be changed. My config files are in the standard locations.
* For cinder (cinder-volume and cinder-backup), glance-api and manila keyrings behavior has changed and Kolla Ansible deployment will not copy those keys using wildcards (ceph.*), instead will use newly introduced variables. Your environment may render unusable after an upgrade if your keys in /etc/kolla/config do not match default values for introduced variables.
* The default behavior for generating the cinder.conf template has changed. An rbd-1 section will be generated when external Ceph functionality is used, i.e. cinder_backend_ceph is set to true. Previously it was only included when Kolla Ansible internal Ceph deployment mechanism was used.
* The rbd section of nova.conf for nova-compute is now generated when nova_backend is set to "rbd". Previously it was only generated when both enable_ceph was "yes" and nova_backend was set to "rbd".
My ceph keys have the default name and are in the default locations. I have cinder_
I added nova_backend: "rbd" and redeployed and now I get a different error: rados.Rados.
http://
What I expected to happen: VMs build without errors after upgrade
How to reproduce it:
Install kolla-ansible Train on Centos 7. Upgrade Centos 7 to 8, upgrade Train to Ussuri, then build a VM.
Environment:
[root@chrnc-
CentOS Linux release 8.2.2004 (Core)
Derived from Red Hat Enterprise Linux 8.2 (Source)
NAME="CentOS Linux"
VERSION="8 (Core)"
ID="centos"
ID_LIKE="rhel fedora"
VERSION_ID="8"
PLATFORM_
PRETTY_NAME="CentOS Linux 8 (Core)"
ANSI_COLOR="0;31"
CPE_NAME=
HOME_URL="https:/
BUG_REPORT_URL="https:/
CENTOS_
CENTOS_
REDHAT_
REDHAT_
CentOS Linux release 8.2.2004 (Core)
CentOS Linux release 8.2.2004 (Core)
cpe:/o:
[root@chrnc-
Linux chrnc-void-
[root@chrnc-
Client: Docker Engine - Community
Version: 20.10.6
API version: 1.41
Go version: go1.13.15
Git commit: 370c289
Built: Fri Apr 9 22:45:33 2021
OS/Arch: linux/amd64
Context: default
Experimental: true
Server: Docker Engine - Community
Engine:
Version: 20.10.6
API version: 1.41 (minimum version 1.12)
Go version: go1.13.15
Git commit: 8728dd2
Built: Fri Apr 9 22:43:57 2021
OS/Arch: linux/amd64
Experimental: false
containerd:
Version: 1.4.4
GitCommit: 05f951a3781f4f2
runc:
Version: 1.0.0-rc93
GitCommit: 12644e614e25b05
docker-init:
Version: 0.19.0
GitCommit: de40ad0
Kolla-Ansible version: Ussuri
kolla_install_type: "source"
Official images
(openstack) [root@chrnc-
---
config_strategy: "COPY_ALWAYS"
virtualenv: /opt/kolla/venv
virtualenv_
kolla_base_distro: "centos"
kolla_install_type: "source"
openstack_release: "train"
node_custom_config: "/etc/kolla/config"
kolla_internal_
network_interface: "eth0"
kolla_external_
neutron_
neutron_
keepalived_
kolla_enable_
openstack_
multiple_
enable_
rabbitmq_
elasticsearch_
kibana_use_v6: true
enable_
enable_ceph: "no"
enable_chrony: "no"
enable_cinder: "yes"
enable_
enable_fluentd: "yes"
enable_grafana: "no"
enable_mariabackup: "yes"
enable_masakari: "yes"
enable_
enable_
enable_
enable_prometheus: "yes"
rabbitmq_
external_
glance_
glance_
glance_
glance_
cinder_
cinder_
nova_backend_ceph: "yes"
nova_compute_
nova_safety_
(openstack) [root@chrnc-
# Please do not change this file directly since it is managed by Ansible and will be overwritten
[global]
cluster network = 172.16.0.0/22
fsid = 30c52736-
mon host = [v2:172.
mon initial members = chrnc-void-
osd pool default crush rule = -1
public network = 172.16.0.0/22
[osd]
osd memory target = 11037206118
(openstack) [root@chrnc-
cluster:
id: 30c52736-
health: HEALTH_OK
services:
mon: 3 daemons, quorum chrnc-void-
mgr: chrnc-void-
osd: 3 osds: 3 up (since 4w), 3 in (since 4w)
data:
pools: 3 pools, 48 pgs
objects: 2.05k objects, 2.9 GiB
usage: 6.1 GiB used, 24 GiB / 30 GiB avail
pgs: 48 active+clean
Changed in kolla-ansible: | |
status: | New → In Progress |
Hi Albert. To verify that the keys in the containers are valid you could docker exec into nova_compute & cinder_volume, and try running some ceph commands using each of the keys.