rabbitMQ became broken after 1 controller node was deleted and 1 added at the same time
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
Fuel for OpenStack |
Confirmed
|
High
|
Vladimir Kuklin | ||
5.1.x |
Won't Fix
|
High
|
Fuel Library (Deprecated) |
Bug Description
On CentOS in HA mode on vcenter's machine,on primary controller deploy of openstack is crashed after addition 1 secondary controller and deletion of another due to reason that rabbitmq can't connect to
primary controller
=======
export VCENTER_
export <email address hidden>'
export VCENTER_
export VCENTER_
=======
Configuration:
=======
steps to reproduce:
1.set up lab on vcenter's machine from 5.1-9(RC3) iso
2.create env and start deploy:
OS: CentOS (HA mode)
roles: 2 controllers,
1 controller + 1 cinder (vmdk)
3. check that deployment of openstack is finished sucessfully
сheck that services on primary controller are available
4. add one secondary controller and delete another secondary controller
5. re-deploy again
6. check that deploy of openstack on primary controller is failed by error:
(/Stage[
Error: unable to connect to node 'rabbit@node-2': nodedown
node-2 is primary controller. And it is available.
Expected result: Deployment process of openstack on each of nodes will be finished successfully
Actual result: deployment of openstack on primary controller is failed due to reason:
(/Stage[
Error: unable to connect to node 'rabbit@node-2': nodedown
-------
-------
[root@nailgun ~]# fuel --fuel-version
api: '1.0'
astute_sha: f5fbd89d1e0e1f2
auth_required: true
build_id: 2014-09-17_04-49-39
build_number: '9'
feature_groups:
- mirantis
fuellib_sha: d9b16846e54f76c
fuelmain_sha: 8ef433e939425ea
nailgun_sha: 51231834c61920a
ostf_sha: 64cb59c681658a7
production: docker
release: '5.1'
release_versions:
2014.1.1-5.1:
VERSION:
api: '1.0'
astute_sha: f5fbd89d1e0e1f2
build_id: 2014-09-17_04-49-39
build_number: '9'
feature_
- mirantis
fuellib_sha: d9b16846e54f76c
fuelmain_sha: 8ef433e939425ea
nailgun_sha: 51231834c61920a
ostf_sha: 64cb59c681658a7
production: docker
release: '5.1'
----
[root@node-2 rabbitmq]# vim <email address hidden>
=INFO REPORT==== 17-Sep-
Error description:
{error,
Log files (may contain more information):
/<email address hidden>
/<email address hidden>
[root@node-2 rabbitmq]# vim <email address hidden>
Stack trace:
[{rabbit_
{rabbit,
{rabbit,
{rpc,
Changed in fuel: | |
assignee: | nobody → Fuel Partner Integration Team (fuel-partner) |
importance: | Undecided → High |
Changed in fuel: | |
assignee: | Fuel Partner Integration Team (fuel-partner) → Fuel Library Team (fuel-library) |
Changed in fuel: | |
status: | New → Won't Fix |
no longer affects: | fuel/5.1.x |
tags: | removed: vcenter |
I am not sure that such a use case is supported right now - we need to investigate how rabbitmq operates when you reset the cluster completely.
Also, it is not clear if you were deleting and adding controllers simultaneously. It looks like bad things can happen if we try to shake rabbitmq cluster in such a way.
I will mark this bug as a known issue for 5.1 and target it to 6.0 for further investigation