Nodes aren't bootstrapped after stopping deployment on provisioning
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
Fuel for OpenStack |
Confirmed
|
Medium
|
Vladimir Sharshov |
Bug Description
Reproduced on {"build_id": "2014-05-
Steps:
1. Create enviroment with all default values
2. Add 1 controller, 1 compute, 1 cinder node
3. Click deploy changes button
4. Wait until provisioning starts
5. Click stop deployment
Expected - nodes went offline and bootstrapped
Actual - nodes went offline and stuck in provisioning state. Nodes are bootstrapped only after force reboot
Diagnostic snapshot is attached and logs from master are also attached
Changed in fuel: | |
assignee: | nobody → Vladimir Sharshov (vsharshov) |
I discovered logs and found that second command which should reboot command failed.
2014-05-06T10:50:42 debug: [395] Run shell command ' echo "Run node rebooting command using 'SB' to sysrq-trigger" kernel/ panic_on_ oops kernel/ panic 0f54-4243- a711-0df1f2768e 5a: Running shell command on nodes ["20", "19", "21"] finished with errors. Nodes [{"uid"=>"20"}, {"uid"=>"19"}, {"uid"=>"21"}] are inaccessible
echo "1" > /proc/sys/
echo "10" > /proc/sys/
echo "b" > /proc/sysrq-trigger
' using ssh
2014-05-06T10:50:42 debug: [395] Run shell command using ssh. Retry 0
2014-05-06T10:50:42 debug: [395] Affected nodes: ["node-20", "node-19", "node-21"]
2014-05-06T10:50:42 debug: [395] Retry result: success nodes: [], error nodes: [], inaccessible nodes: ["node-20", "node-19", "node-21"]
2014-05-06T10:51:12 warning: [395] 7403a914-
Rarely erasing nodes can show such behavior. At now moment i could not reproduce such case. What can we do:
- show error for user, if nodes could not reboot;
- try to reproduce and find why nodes was inaccessible.