train fails on neutron_tempest_plugin.scenario.admin.test_floatingip.FloatingIpTestCasesAdmin.test_two_vms_fips

Bug #1921918 reported by wes hayutin
6
This bug affects 1 person
Affects Status Importance Assigned to Milestone
tripleo
Fix Released
Critical
Unassigned

Bug Description

neutron_tempest_plugin.scenario.admin.test_floatingip.FloatingIpTestCasesAdmin.test_two_vms_fips [22.075122s] ... FAILED

https://logserver.rdoproject.org/openstack-periodic-integration-stable3/opendev.org/openstack/tripleo-ci/master/periodic-tripleo-ci-centos-8-ovb-1ctlr_2comp-featureset020-train/5adb21f/logs/undercloud/var/log/tempest/tempest_run.log.txt.gz

Captured traceback:
~~~~~~~~~~~~~~~~~~~
    Traceback (most recent call last):
      File "/usr/lib/python3.6/site-packages/neutron_tempest_plugin/scenario/admin/test_floatingip.py", line 102, in test_two_vms_fips
        server_ssh_clients, fips = self._create_vms(hyper, avail_zone)
      File "/usr/lib/python3.6/site-packages/neutron_tempest_plugin/scenario/admin/test_floatingip.py", line 77, in _create_vms
        const.SERVER_STATUS_ACTIVE)
      File "/usr/lib/python3.6/site-packages/tempest/common/waiters.py", line 76, in wait_for_server_status
        server_id=server_id)
    tempest.exceptions.BuildErrorException: Server 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c failed to build and is in ERROR status
    Details: {'code': 500, 'created': '2021-03-30T07:29:25Z', 'message': 'Connection to the hypervisor is broken on host', 'details': 'Traceback (most recent call last):\n File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2670, in _build_resources\n yield resources\n File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2444, in _build_and_run_instance\n block_device_info=block_device_info)\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", line 3557, in spawn\n cleanup_instance_disks=created_disks)\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", line 6388, in _create_domain_and_network\n guest.resume()\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/guest.py", line 195, in resume\n self._domain.resume()\n File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 190, in doit\n result = proxy_call(self._autowrap, f, *args, **kwargs)\n File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 148, in proxy_call\n rv = execute(f, *args, **kwargs)\n File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 129, in execute\n six.reraise(c, e, tb)\n File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise\n raise value\n File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 83, in tworker\n rv = meth(*args, **kwargs)\n File "/usr/lib64/python3.6/site-packages/libvirt.py", line 2180, in resume\n if ret == -1: raise libvirtError (\'virDomainResume() failed\', dom=self)\nlibvirt.libvirtError: internal error: client socket is closed\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 503, in get_connection\n conn = self._get_connection()\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 486, in _get_connection\n {\'msg\': ex})\n File "/usr/lib/python3.6/site-packages/oslo_utils/excutils.py", line 220, in __exit__\n self.force_reraise()\n File "/usr/lib/python3.6/site-packages/oslo_utils/excutils.py", line 196, in force_reraise\n six.reraise(self.type_, self.value, self.tb)\n File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise\n raise value\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 475, in _get_connection\n self._wrapped_conn = self._get_new_connection()\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 429, in _get_new_connection\n wrapped_conn = self._connect(self._uri, self._read_only)\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 285, in _connect\n return self._libvirt_proxy.openAuth(uri, auth, flags)\n File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 190, in doit\n result = proxy_call(self._autowrap, f, *args, **kwargs)\n File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 148, in proxy_call\n rv = execute(f, *args, **kwargs)\n File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 129, in execute\n six.reraise(c, e, tb)\n File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise\n raise value\n File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 83, in tworker\n rv = meth(*args, **kwargs)\n File "/usr/lib64/python3.6/site-packages/libvirt.py", line 104, in openAuth\n if ret is None:raise libvirtError(\'virConnectOpenAuth() failed\')\nlibvirt.libvirtError: Failed to connect socket to \'/var/run/libvirt/libvirt-sock\': Connection refused\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2689, in _build_resources\n try_deallocate_networks=deallocate_networks)\n File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2868, in _shutdown_instance\n requested_networks)\n File "/usr/lib/python3.6/site-packages/oslo_utils/excutils.py", line 220, in __exit__\n self.force_reraise()\n File "/usr/lib/python3.6/site-packages/oslo_utils/excutils.py", line 196, in force_reraise\n six.reraise(self.type_, self.value, self.tb)\n File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise\n raise value\n File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2855, in _shutdown_instance\n block_device_info)\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", line 1328, in destroy\n self._destroy(instance)\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", line 1182, in _destroy\n guest = self._host.get_guest(instance)\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 589, in get_guest\n return libvirt_guest.Guest(self._get_domain(instance))\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 604, in _get_domain\n conn = self.get_connection()\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 515, in get_connection\n raise exception.HypervisorUnavailable()\nnova.exception.HypervisorUnavailable: Connection to the hypervisor is broken on host\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2447, in _build_and_run_instance\n instance=instance)\n File "/usr/lib64/python3.6/contextlib.py", line 99, in __exit__\n self.gen.throw(type, value, traceback)\n File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2697, in _build_resources\n reason=six.text_type(exc))\nnova.exception.BuildAbortException: Build of instance 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c aborted: internal error: client socket is closed\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2447, in _build_and_run_instance\n instance=instance)\n File "/usr/lib/python3.6/site-packages/nova/compute/claims.py", line 43, in __exit__\n self.abort()\n File "/usr/lib/python3.6/site-packages/nova/compute/claims.py", line 84, in abort\n self.nodename)\n File "/usr/lib/python3.6/site-packages/oslo_concurrency/lockutils.py", line 328, in inner\n return f(*args, **kwargs)\n File "/usr/lib/python3.6/site-packages/nova/compute/resource_tracker.py", line 571, in abort_instance_claim\n self._update(context.elevated(), self.compute_nodes[nodename])\n File "/usr/lib/python3.6/site-packages/nova/compute/resource_tracker.py", line 1237, in _update\n self._update_to_placement(context, compute_node, startup)\n File "/usr/lib/python3.6/site-packages/retrying.py", line 68, in wrapped_f\n return Retrying(*dargs, **dkw).call(f, *args, **kw)\n File "/usr/lib/python3.6/site-packages/retrying.py", line 223, in call\n return attempt.get(self._wrap_exception)\n File "/usr/lib/python3.6/site-packages/retrying.py", line 261, in get\n six.reraise(self.value[0], self.value[1], self.value[2])\n File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise\n raise value\n File "/usr/lib/python3.6/site-packages/retrying.py", line 217, in call\n attempt = Attempt(fn(*args, **kwargs), attempt_number, False)\n File "/usr/lib/python3.6/site-packages/nova/compute/resource_tracker.py", line 1157, in _update_to_placement\n self.driver.update_provider_tree(prov_tree, nodename)\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", line 7362, in update_provider_tree\n memory_mb = int(self._host.get_memory_mb_total())\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 1061, in get_memory_mb_total\n return self._get_hardware_info()[1]\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 1051, in _get_hardware_info\n return self.get_connection().getInfo()\n File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/host.py", line 515, in get_connection\n raise exception.HypervisorUnavailable()\nnova.exception.HypervisorUnavailable: Connection to the hypervisor is broken on host\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2168, in _do_build_and_run_instance\n filter_properties, request_spec)\n File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2544, in _build_and_run_instance\n instance_uuid=instance.uuid, reason=six.text_type(e))\nnova.exception.RescheduledException: Build of instance 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c was re-scheduled: Connection to the hypervisor is broken on host\n'}

Revision history for this message
Slawek Kaplonski (slaweq) wrote :
Download full text (4.5 KiB)

According to the log from nova-compute https://logserver.rdoproject.org/openstack-periodic-integration-stable3/opendev.org/openstack/tripleo-ci/master/periodic-tripleo-ci-centos-8-ovb-1ctlr_2comp-featureset020-train/5adb21f/logs/overcloud-novacompute-0/var/log/containers/nova/nova-compute.log.txt.gz it seems that this is libvirt issue, not networking:

2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [req-b0dc5c97-2efd-4c17-a144-c08de3dbf812 ff0723602f7e429380f47918acf28178 76617dcd27d64c7d88253e90347937b4 - default default] [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] Instance failed to spawn: libvirt.libvirtError: internal error: client socket is closed
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] Traceback (most recent call last):
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2670, in _build_resources
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] yield resources
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] File "/usr/lib/python3.6/site-packages/nova/compute/manager.py", line 2444, in _build_and_run_instance
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] block_device_info=block_device_info)
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", line 3557, in spawn
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] cleanup_instance_disks=created_disks)
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", line 6388, in _create_domain_and_network
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] guest.resume()
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] File "/usr/lib/python3.6/site-packages/nova/virt/libvirt/guest.py", line 195, in resume
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] self._domain.resume()
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 190, in doit
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] result = proxy_call(self._autowrap, f, *args, **kwargs)
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 148, in proxy_call
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [instance: 25ddccbf-706d-41fa-a6a6-f5f9f70c5b2c] rv = execute(f, *args, **kwargs)
2021-03-30 07:29:22.751 7 ERROR nova.compute.manager [inst...

Read more...

Revision history for this message
Slawek Kaplonski (slaweq) wrote :
Download full text (4.7 KiB)

And journal log from the same node:

Mar 30 07:29:22 overcloud-novacompute-0 systemd-coredump[47824]: Process 24379 (libvirtd) of user 0 dumped core.

                                                                 Stack trace of thread 47803:
                                                                 #0 0x00007faa2f57b9a3 g_source_iter_next (libglib-2.0.so.0)
                                                                 #1 0x00007faa2f57df23 g_main_context_prepare (libglib-2.0.so.0)
                                                                 #2 0x00007faa2f57e9eb g_main_context_iterate.isra.21 (libglib-2.0.so.0)
                                                                 #3 0x00007faa2f57ee72 g_main_loop_run (libglib-2.0.so.0)
                                                                 #4 0x00007faa3291749e n/a (/usr/lib64/libvirt.so.0.6006.0)
Mar 30 07:29:22 overcloud-novacompute-0 systemd[1]: systemd-coredump@0-47823-0.service: Succeeded.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-7d2438f357b22ff8428315270a8745e4b7ccad220f7af5aa3f8ceab9dcdfb3f0-userdata-shm.mount: Succeeded.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[6106]: var-lib-containers-storage-overlay\x2dcontainers-7d2438f357b22ff8428315270a8745e4b7ccad220f7af5aa3f8ceab9dcdfb3f0-userdata-shm.mount: Succeeded.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[1]: var-lib-containers-storage-overlay-33a21bf2b9cc633d874adafcc826e6fb403b68642ab35a57a17fcdfbe7685408-merged.mount: Succeeded.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[6106]: var-lib-containers-storage-overlay-33a21bf2b9cc633d874adafcc826e6fb403b68642ab35a57a17fcdfbe7685408-merged.mount: Succeeded.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[1]: tripleo_nova_libvirt.service: Main process exited, code=exited, status=139/n/a
Mar 30 07:29:22 overcloud-novacompute-0 ovs-vsctl[47901]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --timeout=120 --oneline --format=json --db=tcp:127.0.0.1:6640 -- --if-exists del-port br-int tapb463e039-b2
Mar 30 07:29:22 overcloud-novacompute-0 kernel: device tapb463e039-b2 left promiscuous mode
Mar 30 07:29:22 overcloud-novacompute-0 NetworkManager[1025]: <info> [1617089362.9787] device (tapb463e039-b2): state change: disconnected -> unmanaged (reason 'connection-assumed', sys-iface-state: 'external')
Mar 30 07:29:23 overcloud-novacompute-0 sudo[47906]: neutron : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/usr/bin/neutron-rootwrap /etc/neutron/rootwrap.conf haproxy-kill 9 47781
Mar 30 07:29:23 overcloud-novacompute-0 sudo[47906]: pam_unix(sudo:session): session opened for user root by (uid=0)
Mar 30 07:29:23 overcloud-novacompute-0 dbus-daemon[945]: [system] Activating service name='org.fedoraproject.Setroubleshootd' requested by ':1.56' (uid=0 pid=895 comm="/usr/sbin/sedispatch " label="system_u:system_r:auditd_t:s0") (using servicehelper)
Mar 30 07:29:23 overcloud-novacompute-0 dbus-daemon[47909]: [system] Failed to reset fd limit before activating service: org.freedesktop.DBus.Error.AccessDenied: Failed to restore old fd limit: Operation...

Read more...

Revision history for this message
Martin Schuppert (mschuppert) wrote :

adding the previous segfault message for completeness:

Mar 30 07:29:22 overcloud-novacompute-0 kernel: vm-instance-000[47803]: segfault at 100000038 ip 00007faa2f57b9a3 sp 00007fa9e391d8c0 error 6 in libglib-2.0.so.0.5600.4[7faa2f531000+116000]
Mar 30 07:29:22 overcloud-novacompute-0 kernel: Code: 48 8b 58 08 48 89 5d 10 48 85 db 74 5b 48 8b 03 48 8b 18 48 85 ff 75 1f 48 89 5d 18 31 c0 48 85 db 74 c7 8b 4d 08 85 c9 74 b4 <83> 43 18 01 eb ae 0f 1f 80
00 00 00 00 8b 75 08 85 f6 74 da 8b 47
Mar 30 07:29:22 overcloud-novacompute-0 systemd[1]: Created slice system-systemd\x2dcoredump.slice.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[1]: Started Process Core Dump (PID 47823/UID 0).
Mar 30 07:29:22 overcloud-novacompute-0 ovs-vsctl[47825]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --timeout=120 --oneline --format=json --db=tcp:127.0.0.1:6640 -- --if-exists del-port br-int tap0dbdc767-fd
Mar 30 07:29:22 overcloud-novacompute-0 kernel: device tap0dbdc767-fd left promiscuous mode
Mar 30 07:29:22 overcloud-novacompute-0 NetworkManager[1025]: <info> [1617089362.5930] device (tap0dbdc767-fd): state change: disconnected -> unmanaged (reason 'connection-assumed', sys-iface-state: 'external
')
Mar 30 07:29:22 overcloud-novacompute-0 systemd-coredump[47824]: Process 24379 (libvirtd) of user 0 dumped core.

                                                                 Stack trace of thread 47803:
                                                                 #0 0x00007faa2f57b9a3 g_source_iter_next (libglib-2.0.so.0)
                                                                 #1 0x00007faa2f57df23 g_main_context_prepare (libglib-2.0.so.0)
                                                                 #2 0x00007faa2f57e9eb g_main_context_iterate.isra.21 (libglib-2.0.so.0)
                                                                 #3 0x00007faa2f57ee72 g_main_loop_run (libglib-2.0.so.0)
                                                                 #4 0x00007faa3291749e n/a (/usr/lib64/libvirt.so.0.6006.0)
Mar 30 07:29:22 overcloud-novacompute-0 systemd[1]: systemd-coredump@0-47823-0.service: Succeeded.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-7d2438f357b22ff8428315270a8745e4b7ccad220f7af5aa3f8ceab9dcdfb3f0-userdata-shm.mount: Succeeded.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[6106]: var-lib-containers-storage-overlay\x2dcontainers-7d2438f357b22ff8428315270a8745e4b7ccad220f7af5aa3f8ceab9dcdfb3f0-userdata-shm.mount: Succeeded.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[1]: var-lib-containers-storage-overlay-33a21bf2b9cc633d874adafcc826e6fb403b68642ab35a57a17fcdfbe7685408-merged.mount: Succeeded.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[6106]: var-lib-containers-storage-overlay-33a21bf2b9cc633d874adafcc826e6fb403b68642ab35a57a17fcdfbe7685408-merged.mount: Succeeded.
Mar 30 07:29:22 overcloud-novacompute-0 systemd[1]: tripleo_nova_libvirt.service: Main process exited, code=exited, status=139/n/a

Revision history for this message
Kashyap Chamarthy (kashyapc) wrote :

From a quick look of the libvirtd and liblib coredump / stacktrace, it looks like the following bug. The fix for which was made very recently (early March).

https://bugzilla.redhat.com/show_bug.cgi?id=1931331#c5
https://bugzilla.redhat.com/show_bug.cgi?id=1942010

Upstream libvirt fix for the above is available here:

    https://libvirt.org/git/?p=libvirt.git;a=commit;h=2a490ce5a03
    glib: Use safe glib event workaround in other event loops

The above commit will be part of libvirt v7.0 which will be released on 01 April (yes, really).

Revision history for this message
Ronelle Landy (rlandy) wrote :
Changed in tripleo:
status: Triaged → Fix Released
To post a comment you must log in.
This report contains Public information  
Everyone can see this information.

Other bug subscribers

Remote bug watches

Bug watches keep track of this bug in other bug trackers.