Live Migration failure: Unable to pre-create chardev file console.log: No such file or directory
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
Mirantis OpenStack |
Invalid
|
Medium
|
MOS Nova |
Bug Description
Detailed bug description:
during "boot_attach_
from nova-compute on node-442: http://
As we can see the directory/file was missing
from nova-api on node-1042: http://
Here we can see that the instance tried to be migrated to node-784.
from nova-compute on node-784: http://
Here we can see that nova tried to create the folder and the file
from nova-compute on node-784: http://
And here we can see that we successfully moved the folder later (during deleting process)
root@node-784:~# df -h
Filesystem Size Used Avail Use% Mounted on
udev 2.0G 12K 2.0G 1% /dev
tmpfs 396M 448K 395M 1% /run
/dev/dm-1 141G 2.7G 131G 2% /
none 4.0K 0 4.0K 0% /sys/fs/cgroup
none 5.0M 0 5.0M 0% /run/lock
none 2.0G 0 2.0G 0% /run/shm
none 100M 0 100M 0% /run/user
/dev/vda3 197M 59M 129M 32% /boot
/dev/mapper/vm-nova 53G 33M 53G 1% /var/lib/nova
Steps to reproduce:
1. Deploy Fuel 9.0 from fuel-9.
2. Update Fuel to 9.1:
yum-config-manager --add-repo http://
rpm --import http://
yum install -y python-cudet
yum clean all
update-prepare prepare master
update-prepare update master
3. Applying nailgun patch due to bug https:/
yum -y install patch && curl -s 'https:/
4. fuel rel --sync-
5. Create config for nailgun-agent
mkdir /usr/share/
curl -s 'http://
6. fuel-bootstrap build --verbose --debug --activate --label 'replaced-
7. Create cluster with 3 controllers, 20 osd( both baremetal) and 1000 computes(qemu-kvm)
8. Create additional repo for cluster "mos9.0-proposed":
name: mos9.0-proposed
uri: deb http://
priority: 1200
9. Fix syslog NOFILE limit, due to bug https:/
On fuel node create file /etc/systemd/
[Service]
LimitNOFILE=16384
10. deploy the following cluster:
3 harware controllers, 20 hardware Ceph OSDs, 1000 virtual computes(KVM, 7VMs per hypervisor host),
vxlan+dvr, Ceph for all, OpenStack and deployment debug enabled
11. Applied workaround for https:/
Have changed /etc/libvirt/
migration_port_min = 61152
migration_port_max = 61215
changed iptables rules on each compute node:
sed -i s/"49152:
service iptables-persistent restart
12. Regarding https:/
root@node-549:~# for i in `cat wrong_hyperviso
13. As a workaround for https:/
14. As a workaround for https:/
for i in 1029 1035 162 20 237 259 355 506 541 606 961; do ssh node-$i "virsh secret-set-value a5d0dd94-
for i in 1029 1035 162 20 237 259 355 506 541 606 961; do ssh node-$i "service libvirtd restart"; done
for i in 1029 1035 162 20 237 259 355 506 541 606 961; do ssh node-$i "service nova-compute restart"; done
Expected results:
The test has passed
Actual result:
The test has failed
Reproducibility:
tried once
Workaround:
not yet
Impact:
Live Migration
Description of the environment:
- Operation system: Ubuntu
- Versions of components: MOS 9.1 (9.0-2016-
- Reference architecture: 3 harware controllers, 20 hardware Ceph OSDs, 1000 virtual computes(KVM, 7VMs per hypervisor host), Ceph for all, OpenStack and deployment debug enabled
- Network model: vxlan+dvr
- Related projects installed: -
Additional information:
Diagnostic Snapshot feature doesn't work due https:/
logs from compute node-442: mos-scale-
logs from compute node-442: mos-scale-
logs from controller node-1041: mos-scale-
logs from controller node-1042: mos-scale-
logs from controller node-1043: mos-scale-
description: | updated |
description: | updated |
description: | updated |
description: | updated |
description: | updated |
It was reproduced only on one compute node from one thousand, so let's treat this as Medium until we see it has bigger impact.