Live migration between hosts with provider networks works sporadically
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
networking-ovn |
Fix Released
|
High
|
Richard Theis |
Bug Description
Using this OVS and OVN version:
vagrant@
ovs-vsctl (Open vSwitch) 2.5.90
Compiled Feb 15 2016 03:13:53
DB Schema 7.12.1
vagrant@
With Neutron from master (2-15-2016).
I'm using provider networks for my networking with OVN. When I migrate a VM across a compute host, the networking is sometimes retained, and sometimes it's not. Sometimes it works the first migration, but a subsequent migration fails.
Before a migration, things look like this:
vagrant@
lswitch 6eb40d01-
lport 98cbd16d-
lport 4a0dbb36-
lport b7a641e8-
lswitch cc4c73ae-
lport 1ec64402-
lport provnet-
lswitch 8c23c1ce-
lport provnet-
lport 179edc2a-
lswitch 81a582ea-
lport 583c7fe4-
lport provnet-
lswitch b2b0a19e-
lport ce0c0d02-
lport provnet-
lswitch f8cd658b-
vagrant@
Chassis "0378fa8e-
Encap geneve
ip: "192.168.33.32"
Port_Binding "179edc2a-
Port_Binding "98cbd16d-
Chassis "e4e6d72c-
Encap geneve
ip: "192.168.33.31"
Port_Binding "b7a641e8-
Port_Binding "583c7fe4-
Port_Binding "ce0c0d02-
Chassis "2cc10a56-
Encap geneve
ip: "192.168.33.12"
vagrant@
Post the first migration, they look like this (in a failing condition):
vagrant@
lswitch 6eb40d01-
lport 98cbd16d-
lport 4a0dbb36-
lport b7a641e8-
lswitch cc4c73ae-
lport 1ec64402-
lport provnet-
lswitch 8c23c1ce-
lport provnet-
lport 179edc2a-
lswitch 81a582ea-
lport 583c7fe4-
lport provnet-
lswitch b2b0a19e-
lport ce0c0d02-
lport provnet-
lswitch f8cd658b-
vagrant@
Chassis "0378fa8e-
Encap geneve
ip: "192.168.33.32"
Port_Binding "179edc2a-
Port_Binding "98cbd16d-
Port_Binding "583c7fe4-
Chassis "e4e6d72c-
Encap geneve
ip: "192.168.33.31"
Port_Binding "b7a641e8-
Port_Binding "ce0c0d02-
Chassis "2cc10a56-
Encap geneve
ip: "192.168.33.12"
vagrant@
The VM itself is ok:
agrant@
+------
| ID | Name | Tenant ID | Status | Task State | Power State | Networks |
+------
| 8ba3e11e-
+------
vagrant@
vagrant@
+------
| Property | Value |
+------
| OS-DCF:diskConfig | MANUAL |
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-
| OS-EXT-STS:vm_state | active |
| OS-SRV-
| OS-SRV-
| accessIPv4 | |
| accessIPv6 | |
| config_drive | True |
| created | 2016-02-
| flavor | ovntenant (27) |
| hostId | 1aa531f37cbdfbb
| id | 8ba3e11e-
| image | cirros-
| key_name | - |
| locked | False |
| metadata | {} |
| name | vm1 |
| os-extended-
| progress | 80 |
| provider network | 10.10.0.59 |
| security_groups | default |
| status | ACTIVE |
| tenant_id | e588180c1c9a44c
| updated | 2016-02-
| user_id | 4ead3522ebbf46c
+------
vagrant@
I cannot get to the console however:
vagrant@
vagrant@
Changed in networking-ovn: | |
assignee: | nobody → Richard Theis (rtheis) |
Is there any chance you could provide access to a compute node with a VM in the failed state? If not, maybe we could just get together on IRC for some live debugging.