Yes, I refer to tripleo undercloud vm: (overcloud) [stack@undercloud-0 ~]$ openstack server list +--------------------------------------+------------------------+---------+------------------------------------+--------+--------+ | ID | Name | Status | Networks | Image | Flavor | +--------------------------------------+------------------------+---------+------------------------------------+--------+--------+ | 7716bfde-d441-4c07-9bb8-b290607a6d4e | selfservice-instance4 | SHUTOFF | selfservice3=10.3.0.16 | cirros | cirros | | 521d8243-8255-4746-a8d9-5d5d2035a2a1 | selfservice3-instance2 | ACTIVE | selfservice3=10.3.0.8 | cirros | cirros | | 164dd0d7-cb1e-45ad-abbf-0f8de6dd9f97 | selfservice3-instance | ACTIVE | selfservice3=10.3.0.25 | cirros | cirros | | 9d21306a-4e6d-4250-9936-9f8ac0895911 | selfservice2-instance5 | ACTIVE | selfservice2=10.2.0.6 | cirros | cirros | | 05797371-111d-419b-9199-c787f52f0555 | selfservice2-instance3 | ACTIVE | selfservice2=10.2.0.5 | cirros | cirros | | 6d68a21d-5361-4e37-b0a6-4a9ebfea3e82 | selfservice2-instance2 | ACTIVE | selfservice2=10.2.0.18 | cirros | cirros | | aa3eed0a-473b-4e71-9c33-23e7aef09f44 | selfservice2-instance | ACTIVE | selfservice2=10.2.0.23, 10.0.0.232 | cirros | cirros | | d4ad2ca9-21ec-40e6-897d-5ec48e77765f | selfservice-instance3 | ACTIVE | selfservice=10.1.0.5 | cirros | cirros | | ed68c202-2a57-4078-b459-4e5650adc185 | selfservice-instance2 | ACTIVE | selfservice=10.1.0.20 | cirros | cirros | | ba63ce4f-f5d6-4634-bd40-868a57c4ecfd | selfservice-instance | ACTIVE | selfservice=10.1.0.12, 10.0.0.234 | cirros | cirros | +--------------------------------------+------------------------+---------+------------------------------------+--------+--------+ (overcloud) [stack@undercloud-0 ~]$ (overcloud) [stack@undercloud-0 ~]$ (overcloud) [stack@undercloud-0 ~]$ . stackrc (undercloud) [stack@undercloud-0 ~]$ openstack server list +--------------------------------------+--------------+--------+------------------------+----------------+------------+ | ID | Name | Status | Networks | Image | Flavor | +--------------------------------------+--------------+--------+------------------------+----------------+------------+ | b71c1bab-8e11-4958-bb36-1556c9711e57 | controller-0 | ACTIVE | ctlplane=192.168.24.11 | overcloud-full | controller | | c4a66c10-b213-4aaa-b575-b69872fd8d28 | controller-2 | ACTIVE | ctlplane=192.168.24.21 | overcloud-full | controller | | b0cd4f07-3b9b-4f5f-baea-bbaf8454c6f6 | controller-1 | ACTIVE | ctlplane=192.168.24.7 | overcloud-full | controller | | 97fb6e74-bc9b-4a9e-b369-895fd1401277 | compute-2 | ACTIVE | ctlplane=192.168.24.12 | overcloud-full | compute | | 0c4d8ad5-6c0a-4508-98ed-34c263b98960 | compute-0 | ACTIVE | ctlplane=192.168.24.6 | overcloud-full | compute | | 01a11464-a9ab-4f97-a87f-cf7f1290599c | compute-1 | ACTIVE | ctlplane=192.168.24.9 | overcloud-full | compute | +--------------------------------------+--------------+--------+------------------------+----------------+------------+ (overcloud) [stack@undercloud-0 ~]$ ping 10.0.0.234 PING 10.0.0.234 (10.0.0.234) 56(84) bytes of data. 64 bytes from 10.0.0.234: icmp_seq=1 ttl=62 time=0.512 ms 64 bytes from 10.0.0.234: icmp_seq=2 ttl=62 time=0.525 ms 64 bytes from 10.0.0.234: icmp_seq=3 ttl=62 time=0.453 ms ..... After systemctl restart: (overcloud) [stack@undercloud-0 ~]$ ping 10.0.0.234 PING 10.0.0.234 (10.0.0.234) 56(84) bytes of data. 64 bytes from 10.0.0.234: icmp_seq=1 ttl=62 time=0.512 ms 64 bytes from 10.0.0.234: icmp_seq=2 ttl=62 time=0.525 ms 64 bytes from 10.0.0.234: icmp_seq=3 ttl=62 time=0.453 ms ..... (undercloud) [stack@undercloud-0 ~]$ ping 10.0.0.234 PING 10.0.0.234 (10.0.0.234) 56(84) bytes of data. From 10.0.0.29 icmp_seq=1 Destination Host Unreachable From 10.0.0.29 icmp_seq=2 Destination Host Unreachable From 10.0.0.29 icmp_seq=3 Destination Host Unreachable From 10.0.0.29 icmp_seq=4 Destination Host Unreachable ----------------- Actions in compute2: [root@compute-2 heat-admin]# date;systemctl restart openvswitch Mon Nov 26 08:52:14 UTC 2018 [root@compute-2 heat-admin]# [root@compute-2 heat-admin]# [root@compute-2 heat-admin]# date;docker restart neutron_ovs_agent Mon Nov 26 08:53:19 UTC 2018 neutron_ovs_agent [root@compute-2 heat-admin]# I attach ovs agent logs. as you comment, the problem seems that the flows are not correctly reloaded after the restarts. with problem 2: There are some comunication problem after some kill: root@compute-2 heat-admin]# tail -f /var/log/containers/neutron/metadata-agent.log 2018-11-26 08:51:10.883 8224 INFO eventlet.wsgi.server [-] 10.2.0.18, "GET /2009-04-04/meta-data/placement/availability-zone HTTP/1.1" status: 200 len: 139 time: 0.1934390 2018-11-26 08:51:10.968 8223 INFO eventlet.wsgi.server [-] 10.3.0.25, "GET /2009-04-04/meta-data/placement/availability-zone HTTP/1.1" status: 200 len: 139 time: 0.1648369 2018-11-26 09:11:20.883 8224 WARNING oslo.messaging._drivers.impl_rabbit [-] Unexpected error during heartbeart thread processing, retrying...: IOError: Socket closed 2018-11-26 09:11:34.307 7137 ERROR oslo.messaging._drivers.impl_rabbit [-] [357cec88-105c-4f2b-b26c-70ae3864ec0f] AMQP server controller-2.internalapi.localdomain:5672 closed the connection. Check login credentials: Socket closed: IOError: Socket closed 2018-11-26 09:11:35.323 7137 INFO oslo.messaging._drivers.impl_rabbit [-] [357cec88-105c-4f2b-b26c-70ae3864ec0f] Reconnected to AMQP server on controller-2.internalapi.localdomain:5672 via [amqp] client with port 50096. 2018-11-26 09:11:43.938 8223 ERROR oslo.messaging._drivers.impl_rabbit [-] [24405812-662e-4c40-b834-37a07d80366f] AMQP server controller-1.internalapi.localdomain:5672 closed the connection. Check login credentials: Socket closed: IOError: Socket closed 2018-11-26 09:11:44.955 8223 INFO oslo.messaging._drivers.impl_rabbit [-] [24405812-662e-4c40-b834-37a07d80366f] Reconnected to AMQP server on controller-1.internalapi.localdomain:5672 via [amqp] client with port 37994. 2018-11-26 09:11:45.218 8224 ERROR oslo.messaging._drivers.impl_rabbit [-] [7bf56d8a-afcf-4b87-b2dd-c865c2faf08f] AMQP server controller-2.internalapi.localdomain:5672 closed the connection. Check login credentials: Socket closed: IOError: Socket closed 2018-11-26 09:11:46.235 8224 INFO oslo.messaging._drivers.impl_rabbit [-] [7bf56d8a-afcf-4b87-b2dd-c865c2faf08f] Reconnected to AMQP server on controller-2.internalapi.localdomain:5672 via [amqp] client with port 50116. 2018-11-26 09:11:59.117 8223 WARNING oslo.messaging._drivers.impl_rabbit [-] Unexpected error during heartbeart thread processing, retrying...: IOError: Socket closed 2018-11-26 09:12:04.417 7137 ERROR oslo.messaging._drivers.impl_rabbit [-] [8f9c38d5-3ad8-4cab-8446-a2f07e6d370f] AMQP server controller-1.internalapi.localdomain:5672 closed the connection. Check login credentials: Socket closed: IOError: Socket closed 2018-11-26 09:12:05.444 7137 INFO oslo.messaging._drivers.impl_rabbit [-] [8f9c38d5-3ad8-4cab-8446-a2f07e6d370f] Reconnected to AMQP server on controller-1.internalapi.localdomain:5672 via [amqp] client with port 38042. 2018-11-26 09:12:15.386 8224 WARNING oslo.messaging._drivers.impl_rabbit [-] Unexpected error during heartbeart thread processing, retrying...: error: [Errno 104] Connection reset by peer .................... vi /var/log/containers/neutron/openvswitch-agent.log 2018-11-26 09:12:55.615 28480 ERROR neutron.plugins.ml2.drivers.openvswitch.agent.openflow.native.ofswitch [req-6fcf8cf0-e4cb-4142-b6ea-16efb2115491 - - - - -] Switch connection timeout 2018-11-26 09:12:55.616 28480 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbGetCommand(column=datapath_id, table=Bridge, record=br-tun) do_commit /usr/lib/python2.7/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:84 2018-11-26 09:12:55.616 28480 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python2.7/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:121 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server [req-6fcf8cf0-e4cb-4142-b6ea-16efb2115491 - - - - -] Exception during message handling: RuntimeError: Switch connection timeout 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server Traceback (most recent call last): 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/oslo_messaging/rpc/server.py", line 163, in _process_incoming 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/oslo_messaging/rpc/dispatcher.py", line 265, in dispatch 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/oslo_messaging/rpc/dispatcher.py", line 194, in _do_dispatch 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/osprofiler/profiler.py", line 159, in wrapper 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server result = f(*args, **kwargs) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/oslo_log/helpers.py", line 67, in wrapper 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server return method(*args, **kwargs) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/neutron/plugins/ml2/drivers/l2pop/rpc_manager/l2population_rpc.py", line 41, in add_fdb_entries 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server self.fdb_add(context, self._unmarshall_fdb_entries(fdb_entries)) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/osprofiler/profiler.py", line 159, in wrapper 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server result = f(*args, **kwargs) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/neutron/plugins/ml2/drivers/openvswitch/agent/ovs_neutron_agent.py", line 541, in fdb_add 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server agent_ports, self._tunnel_port_lookup) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/osprofiler/profiler.py", line 159, in wrapper 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server result = f(*args, **kwargs) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/oslo_log/helpers.py", line 67, in wrapper 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server return method(*args, **kwargs) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/neutron/plugins/ml2/drivers/l2pop/rpc_manager/l2population_rpc.py", line 254, in fdb_add_tun 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server self.add_fdb_flow(br, port, remote_ip, lvm, ofport) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/osprofiler/profiler.py", line 159, in wrapper 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server result = f(*args, **kwargs) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/neutron/plugins/ml2/drivers/openvswitch/agent/ovs_neutron_agent.py", line 561, in add_fdb_flow 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server lvm.tun_ofports) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/neutron/plugins/ml2/drivers/openvswitch/agent/openflow/native/br_tun.py", line 174, in install_flood_to_tun 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server (_dp, ofp, ofpp) = self._get_dp() 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/neutron/plugins/ml2/drivers/openvswitch/agent/openflow/native/ovs_bridge.py", line 69, in _get_dp 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server self._cached_dpid = new_dpid 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 220, in __exit__ 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server self.force_reraise() 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 196, in force_reraise 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server six.reraise(self.type_, self.value, self.tb) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/neutron/plugins/ml2/drivers/openvswitch/agent/openflow/native/ovs_bridge.py", line 52, in _get_dp 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server dp = self._get_dp_by_dpid(self._cached_dpid) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server File "/usr/lib/python2.7/site-packages/neutron/plugins/ml2/drivers/openvswitch/agent/openflow/native/ofswitch.py", line 79, in _get_dp_by_dpid 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server raise RuntimeError(m) 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server RuntimeError: Switch connection timeout 2018-11-26 09:12:55.617 28480 ERROR oslo_messaging.rpc.server 2018-11-26 09:12:56.272 28480 ERROR neutron.plugins.ml2.drivers.openvswitch.agent.openflow.native.ofswitch [req-6fcf8cf0-e4cb-4142-b6ea-16efb2115491 - - - - -] Switch connection timeout 2018-11-26 09:12:56.273 28480 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbGetCommand(column=datapath_id, table=Bridge, record=br-tun) do_commit /usr/lib/python2.7/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:84