Instance failed to spawn after controllers reboot
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
Mirantis OpenStack |
Fix Released
|
High
|
Dmitry Mescheryakov | ||
7.0.x |
Fix Released
|
High
|
Dmitry Mescheryakov | ||
8.0.x |
Fix Released
|
High
|
Dmitry Mescheryakov | ||
9.x |
Fix Released
|
High
|
Dmitry Mescheryakov |
Bug Description
7.0 MU1
1. Revert snapshot 3 ['controller', 'mongo'] +2 ['compute', 'cinder']
2. reboot --force controller
3. Wait until controller is switching in maintenance mode
4. Exit maintenance mode
5. Check the controller become available
6. Run OSTF and repeat for other 2 nodes.
For the 3rd node OSTF smoke finally didn't pass
{
"Create volume and boot instance from it (failure)": "Failed to get to expected status. In error state. Please refer to OpenStack logs for more details."
},
{
"Create volume and attach it to instance (failure)": "Failed to get to expected status. In error state. Please refer to OpenStack logs for more details."
},
{
"Check network connectivity from instance via floating IP (failure)": "Failed to get to expected status. In error state. Please refer to OpenStack logs for more details."
},
{
"Launch instance (failure)": "Failed to get to expected status. In error state. Please refer to OpenStack logs for more details."
},
{
"Launch instance with file injection (failure)": "Failed to get to expected status. In error state. Please refer to OpenStack logs for more details."
},
{
"Launch instance, create snapshot, launch instance from snapshot (failure)": "Failed to get to expected status. In error state. Please refer to OpenStack logs for more details."
}
]
Errors in nova-compute "NovaException: Unexpected vif_type=
https:/
Open vSwitch agents are marked as dead in neutron agent-list the same as others:
root@node-5:~# neutron agent-list
+------
| id | agent_type | host | alive | admin_state_up | binary |
+------
| 122b0a93-
| 3b1e0177-
| 49b5dac6-
| 4b9259b6-
| 577098e7-
| 70b2841c-
| 992c731e-
| 99eabcd2-
| a066f867-
| c6028891-
| ccc0bcb2-
| d84f6b13-
| f923b54d-
| fbd47cae-
+------
Workaround: Kill neutron-
summary: |
- Instance failed to spawn after controller's reboot + Instance failed to spawn after controllers reboot |
Changed in mos: | |
status: | New → Confirmed |
Changed in mos: | |
assignee: | MOS Neutron (mos-neutron) → Andrey Epifanov (aepifanov) |
Changed in mos: | |
importance: | Critical → High |
Changed in mos: | |
status: | Confirmed → In Progress |
Changed in mos: | |
assignee: | Andrey Epifanov (aepifanov) → Dmitry Mescheryakov (dmitrymex) |
tags: |
added: area-oslo hit-hcf removed: neutron |
tags: | added: on-verification |
Also we can see crush report in rabbit logs during the test execution(https:/ /bugs.launchpad .net/fuel/ +bug/1513511), but the test firstly run HA test and verify rabbit
2015-11-10 12:06:21,821 - INFO fuel_web_ client. py:243 -- OSTF test statuses are : {
"Check if amount of tables in databases is the same on each node": "success",
"RabbitMQ replication": "success",
"Check pacemaker status": "success",
"RabbitMQ availability": "success",
"Check galera environment state": "success",
"Check data replication over mysql": "success"
}
and then start ['smoke', 'sanity']
In reverted snapshot I've seen that rabbitmq was not stopped