When kill(sometines doesn't restart) the ovs switch or restart it in the compute nodes vm conectivity is lost
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
neutron |
Fix Released
|
Medium
|
Slawek Kaplonski |
Bug Description
OSP 14
3 controllers + 3 computes + dvr
several vms in one compute with fip.
Problem 1 :
root@compute-2 heat-admin]# systemctl restart openvswitch
fip conectivity with undercloud vm is lost and no recover
conectivity with other computes is lost, but it is recovered restarting neutron openvswitch agent container
root@compute-2 heat-admin]# systemctl restart openvswitch.
Problem 2:
After kill -9 "pid ovs switch"
Sometimes the ovs switch in not restarted automatically
Same problems that in the scenario1
[root@compute-2 heat-admin]# ps -ef | grep ovs
root 10558 7292 0 12:09 ? 00:00:01 /usr/bin/python2 /bin/privsep-helper --config-file /usr/share/
42435 46886 46871 0 13:17 ? 00:00:00 /bin/bash /neutron_
openvsw+ 49054 1 1 13:20 ? 00:00:00 ovs-vswitchd unix:/var/
root 49217 17666 0 13:21 pts/0 00:00:00 grep --color=auto ovs
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]# ps -ef | grep ovs
root 10558 7292 0 12:09 ? 00:00:01 /usr/bin/python2 /bin/privsep-helper --config-file /usr/share/
42435 46886 46871 0 13:17 ? 00:00:00 /bin/bash /neutron_
openvsw+ 49054 1 0 13:20 ? 00:00:00 ovs-vswitchd unix:/var/
root 49421 17666 0 13:21 pts/0 00:00:00 grep --color=auto ovs
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]# ps -ef | grep ovs
root 10558 7292 0 12:09 ? 00:00:01 /usr/bin/python2 /bin/privsep-helper --config-file /usr/share/
42435 46886 46871 0 13:17 ? 00:00:00 /bin/bash /neutron_
openvsw+ 49054 1 0 13:20 ? 00:00:00 ovs-vswitchd unix:/var/
root 49423 17666 0 13:21 pts/0 00:00:00 grep --color=auto ovs
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]# kill -9 49054
[root@compute-2 heat-admin]# ps -ef | grep ovs
root 10558 7292 0 12:09 ? 00:00:01 /usr/bin/python2 /bin/privsep-helper --config-file /usr/share/
42435 46886 46871 0 13:17 ? 00:00:00 /bin/bash /neutron_
root 49610 17666 0 13:22 pts/0 00:00:00 grep --color=auto ovs
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]# ps -ef | grep ovs
root 10558 7292 0 12:09 ? 00:00:01 /usr/bin/python2 /bin/privsep-helper --config-file /usr/share/
42435 46886 46871 0 13:17 ? 00:00:00 /bin/bash /neutron_
root 49628 17666 0 13:22 pts/0 00:00:00 grep --color=auto ovs
[root@compute-2 heat-admin]# date
Fri Nov 23 13:22:22 UTC 2018
[root@compute-2 heat-admin]# ps -ef | grep ovs
root 10558 7292 0 12:09 ? 00:00:01 /usr/bin/python2 /bin/privsep-helper --config-file /usr/share/
42435 46886 46871 0 13:17 ? 00:00:00 /bin/bash /neutron_
root 49788 17666 0 13:22 pts/0 00:00:00 grep --color=auto ovs
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]#
[root@compute-2 heat-admin]# ps -ef | grep ovs
root 10558 7292 0 12:09 ? 00:00:01 /usr/bin/python2 /bin/privsep-helper --config-file /usr/share/
42435 46886 46871 0 13:17 ? 00:00:00 /bin/bash /neutron_
root 49790 17666 0 13:22 pts/0 00:00:00 grep --color=auto ovs
overcloud) [stack@undercloud-0 ~]$ openstack versions show
+------
| Region Name | Service Type | Version | Status | Endpoint | Min Microversion | Max Microversion |
+------
| regionOne | block-storage | 2.0 | DEPRECATED | http://
| regionOne | block-storage | 3.0 | CURRENT | http://
| regionOne | placement | None | CURRENT | http://
| regionOne | network | 2.0 | CURRENT | http://
| regionOne | alarm | 2.0 | CURRENT | http://
| regionOne | cloudformation | 1.0 | CURRENT | http://
| regionOne | event | 2.0 | CURRENT | http://
| regionOne | orchestration | 1.0 | CURRENT | http://
| regionOne | object-store | 1.0 | CURRENT | http://
| regionOne | compute | 2.0 | SUPPORTED | http://
| regionOne | compute | 2.1 | CURRENT | http://
| regionOne | image | 2.0 | SUPPORTED | http://
| regionOne | image | 2.1 | SUPPORTED | http://
| regionOne | image | 2.2 | SUPPORTED | http://
| regionOne | image | 2.3 | SUPPORTED | http://
| regionOne | image | 2.4 | SUPPORTED | http://
| regionOne | image | 2.5 | SUPPORTED | http://
| regionOne | image | 2.6 | SUPPORTED | http://
| regionOne | image | 2.7 | CURRENT | http://
| regionOne | metric | 1.0 | CURRENT | http://
| regionOne | identity | 3.10 | CURRENT | http://
+------
(overcloud) [stack@undercloud-0 ~]$ cat /etc/re
redhat-lsb/ redhat-release request-key.conf request-key.d/ resolv.conf
(overcloud) [stack@undercloud-0 ~]$ cat /etc/re
redhat-lsb/ redhat-release request-key.conf request-key.d/ resolv.conf
(overcloud) [stack@undercloud-0 ~]$ cat /etc/redhat-release
Red Hat Enterprise Linux Server release 7.6 (Maipo)
tags: | added: ovs |
Changed in neutron: | |
status: | Incomplete → Confirmed |
importance: | Undecided → Medium |
Changed in neutron: | |
assignee: | nobody → Slawek Kaplonski (slaweq) |
status: | Confirmed → In Progress |
by undercloud/ overcloud, do you mean triple-o?
problem 1 means you rebooted openvswitch (ovs-vswitchd/ ovsdb-server) without restarting
neutron-ovs-agent? maybe canary flow check is not working well for some reasons. can you provide the log of ovs agent?
problem 2 means you killed ovs-vswitchd? looks unrelated to neutron.