Killing "backup" amphora before "master" is recovered leads to the fact that topology is not restored
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
octavia |
Invalid
|
High
|
Unassigned |
Bug Description
Env: ocata packages with octavia
Set
[controller_worker]
loadbalancer_
[house_keeping]
spare_amphora_
in the etc/octavia.conf
Steps to reproduce:
1. Check that 1 amphora VM was created
2. Create an instance
3. SSH to the instance and start two servers
4. Create a load balancer graph with two members and with ROUND_ROBIN algorithm (LB, listener, pool with members)
5. Associate the VIP from public net
6. Check that one more amphora per LB was created and there "master" and "backup" amphoras for LB.
Spare pool was filled again by creation of one more amphora
7. Send NUM requests to the floating ip and check that they are shared
between the two servers
8. Delete amphora with "master" role
9. "Backup" amphora should balance traffic between the two servers
10. Kill "backup" amphora
Expected result:
After some time new "master" amphorae from spare pool is created and new backup amphora too.
Topology is recovered.
Observed result:
After some time new "master" amphorae from spare pool is created, but there is no "backup" amphora.
Deletion of such LB leads to error like "lb is in immutable state".
Neutron and octavia DB have different info about LB state.
Changed in octavia: | |
status: | New → Incomplete |
Changed in octavia: | |
status: | Incomplete → Triaged |
importance: | Undecided → Critical |
Changed in octavia: | |
importance: | Critical → High |
Hello agian,
So the issue with the neutron database getting out of sync is a known issue. We are doing two things to address this issue: /review. openstack. org/#/c/ 478385/
1. We are getting rid of neutron-lbaas so we will no longer have the neutron database to get out of sync.
2. In the interim there is a patch up for review that will put in place a workaround: https:/
Can you provide the octavia database status information? Preferably the amphora record, amphora_health record, load balancer record, and listener record?
What I would expect is the health monitor should come back around and notice that the backup amphora is not healthy and start a rebuild on it as well. Having the octavia database state information would help us debug the situation.