[arm64 ocata] newly created instances are unable to raise network interfaces
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
QEMU |
Fix Released
|
Undecided
|
Unassigned | ||
Ubuntu Cloud Archive |
Fix Released
|
Undecided
|
Unassigned | ||
Ocata |
Fix Committed
|
High
|
Unassigned | ||
libvirt |
New
|
Undecided
|
Unassigned | ||
libvirt (Ubuntu) |
Invalid
|
Undecided
|
Unassigned | ||
qemu (Ubuntu) |
Fix Released
|
Undecided
|
Unassigned | ||
Zesty |
Fix Released
|
Undecided
|
Unassigned |
Bug Description
[Impact]
* A change in qemu 2.8 (83d768b virtio: set ISR on dataplane
notifications) broke virtio handling on platforms without a
controller. Those encounter flaky networking due to missed IRQs
* Fix is a backport of the upstream fix b4b9862b: virtio: Fix no
interrupt when not creating msi controller
[Test Case]
* On Arm with Zesty (or Ocata) run a guest without PCI based devices
* Example in e.g. c#23
* Without the fix the networking does not work reliably (as it losses
IRQs), with the fix it works fine.
[Regression Potential]
* Changing the IRQ handling of virtio could affect virtio in general.
But when reviwing the patch you'll see that it is small and actually
only changes to enable IRQ on one more place. That could cause more
IRQs than needed in the worst case, but those are usually not
breaking but only slowing things down. Also this fix is upstream
quite a while, increasing confidence.
[Other Info]
* There is currently 1720397 in flight in the SRU queue, so acceptance
of this upload has to wait until that completes.
---
arm64 Ocata ,
I'm testing to see I can get Ocata running on arm64 and using the openstack-base bundle to deploy it. I have added the bundle to the log file attached to this bug.
When I create a new instance via nova, the VM comes up and runs, however fails to raise its eth0 interface. This occurs on both internal and external networks.
ubuntu@
+------
| ID | Name | Status | Task State | Power State | Networks |
+------
| dcaf6d51-
| aa0b8aee-
+------
ubuntu@
+------
| Property | Value |
+------
| OS-DCF:diskConfig | MANUAL |
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-STS:vm_state | active |
| OS-SRV-
| OS-SRV-
| accessIPv4 | |
| accessIPv6 | |
| config_drive | |
| created | 2017-09-
| flavor | m1.small (717660ae-
| hostId | 5612a00671c4725
| id | aa0b8aee-
| image | zestynosplash (e88fd1bd-
| internal network | 10.5.5.13 |
| key_name | mykey |
| metadata | {} |
| name | sfeole2 |
| os-extended-
| progress | 0 |
| security_groups | default |
| status | ACTIVE |
| tenant_id | 9f7a21c1ad264fe
| updated | 2017-09-
| user_id | e6bb6f5178a248c
+------
As seen above the instances boot an run. Full Console output is attached to this bug.
[ OK ] Started Initial cloud-init job (pre-networking).
[ OK ] Reached target Network (Pre).
[ OK ] Started AppArmor initialization.
Starting Raise network interfaces...
[FAILED] Failed to start Raise network interfaces.
See 'systemctl status networking.service' for details.
Starting Initial cloud-init job (metadata service crawler)...
[ OK ] Reached target Network.
[ 315.051902] cloud-init[881]: Cloud-init v. 0.7.9 running 'init' at Fri, 22 Sep 2017 18:29:15 +0000. Up 314.70 seconds.
[ 315.057291] cloud-init[881]: ci-info: +++++++
[ 315.060338] cloud-init[881]: ci-info: +------
[ 315.063308] cloud-init[881]: ci-info: | Device | Up | Address | Mask | Scope | Hw-Address |
[ 315.066304] cloud-init[881]: ci-info: +------
[ 315.069303] cloud-init[881]: ci-info: | eth0: | True | . | . | . | fa:16:3e:39:4c:48 |
[ 315.072308] cloud-init[881]: ci-info: | eth0: | True | . | . | d | fa:16:3e:39:4c:48 |
[ 315.075260] cloud-init[881]: ci-info: | lo: | True | 127.0.0.1 | 255.0.0.0 | . | . |
[ 315.078258] cloud-init[881]: ci-info: | lo: | True | . | . | d | . |
[ 315.081249] cloud-init[881]: ci-info: +------
[ 315.084240] cloud-init[881]: 2017-09-22 18:29:15,393 - url_helper.
----------------
I have checked all services in neutron and made sure that they are running and restarted in neutron-gateway
[ + ] neutron-dhcp-agent
[ + ] neutron-l3-agent
[ + ] neutron-
[ + ] neutron-
[ + ] neutron-
[ + ] neutron-
[ + ] neutron-ovs-cleanup
and have also restarted and checked the nova-compute logs for their neutron counterparts.
[ + ] neutron-
[ + ] neutron-ovs-cleanup
There are some warnings/errors in the neutron-gateway logs which I have attached the full tarball in a separate attachment to this bug:
2017-09-24 14:39:33.152 10322 INFO ryu.base.
2017-09-24 14:39:33.153 10322 INFO ryu.base.
2017-09-24 14:39:39.577 10322 ERROR neutron.
These warnings/errors also occur on the nova-compute hosts in /var/log/neutron/
2017-09-22 18:54:52.130 387556 INFO ryu.base.
2017-09-22 18:54:56.124 387556 ERROR neutron.
I'm not sure if the above error could be pertaining to the failure, I have also attached those logs to this bug as well...
.
(neutron) agent-list
+------
| id | agent_type | host | availability_zone | alive | admin_state_up | binary |
+------
| 0cca03fb-
| 14a5fd52-
| 2ebc7238-
| 4f6275be-
| 86ecc6b0-
| 947ad3ab-
| 996b0692-
| ab6b1065-
| fe24f622-
+------
Should neutron-openvswitch be assigned to the 'nova' availability zone?
I have also attached the ovs-vsctl show from a nova-compute host and the neutron-gateway host to ensure that the open v switch routes are correct.
I can supply more logs if required.
no longer affects: | cloud-archive/pike |
Changed in cloud-archive: | |
status: | New → Fix Released |
description: | updated |
Changed in qemu (Ubuntu Zesty): | |
status: | Incomplete → Fix Committed |
tags: | added: verification-needed verification-needed-zesty |
tags: |
added: verification-done-zesty removed: verification-needed-zesty |
juju status
bundle file
vm console output
and nova service list
can be found in ocata-arm64- neutronbug. txt