bootstrap issue with replicasets on 1.20.1 with VM on MAAS provider
| Affects | Status | Importance | Assigned to | Milestone | |
|---|---|---|---|---|---|
| juju-core |
High
|
Unassigned | |||
Bug Description
I'm trying to bootstrap Juju 1.20.1 using a MAAS provider. The node I'm bootstrapping on is a VM on an OrangeBox.
I hit repeated
2014-07-11 10:25:08 WARNING juju.replicaset replicaset.go:86 Initiate: fetching replication status failed: cannot get replica set status: can't get local.system.
Followed by
2014-07-11 10:25:08 INFO juju.worker.
2014-07-11 10:25:08 ERROR juju.cmd supercommand.go:323 cannot initiate replica set: cannot get replica set status: can't get local.system.
ERROR bootstrap failed: subprocess encountered error code 1
Stopping instance...
Bootstrap failed, destroying environment
ERROR subprocess encountered error code 1
Full log from bootstrap on http://
I was asked to grab some mongo logs, but I'm struggling to do that since the bootstrap nodes is torn down. So far this is 100% reproducible.
| description: | updated |
| Adam Collard (adam-collard) wrote : | #1 |
| Adam Collard (adam-collard) wrote : | #2 |
# sudo find . -name '*mongo*'
./etc/default/
./usr/share/
./usr/share/
./usr/lib/
./usr/lib/
./usr/lib/
./usr/lib/
./usr/lib/
./usr/lib/
./tmp/mongodb-
./var/lib/
./var/lib/
./var/lib/
./var/cache/
This looks odd to me, distinct lack of mongo bits?
| Adam Collard (adam-collard) wrote : | #3 |
Found mongo entries in /var/log/syslog. http://
| Adam Collard (adam-collard) wrote : | #4 |
Note that I only see this when bootstrapping to a VM. I don't see it for physical nodes.
| summary: |
- bootstrap issue with replicasets on 1.20.1 with MAAS provider + bootstrap issue with replicasets on 1.20.1 with VM on MAAS provider |
| Adam Collard (adam-collard) wrote : | #5 |
I see it always for this particular VM, but can't reproduce on other VMs in a (almost) parallel setup
| tags: | added: bootstrap mongodb |
| Changed in juju-core: | |
| status: | New → Triaged |
| importance: | Undecided → High |
| milestone: | none → 1.21-alpha1 |
| Michael Foord (mfoord) wrote : | #6 |
From the syslog entries it appears that the real error is that mongodb can't reach the host machine via the address we provide to it when creating (initiating) the replica set. So it's a networking issue. However this error is masked by our bad error reporting. I've raised bug 1340749 to address that problem.
| Ian Booth (wallyworld) wrote : | #7 |
Marking this bug as invalid as the root cause is a networking setup issue external to juju-core.
| Changed in juju-core: | |
| status: | Triaged → Invalid |
| Martin Packman (gz) wrote : | #8 |
The underlying issue here is juju is being given a bogus address via the MAAS api.
On bootstrap, we cope with this, because we attempt to connect to each address in turn, and the 10.14.4.76 one is correct and succeeds:
Attempting to connect to node0vm0.maas:22
Attempting to connect to 10.14.4.59:22
Attempting to connect to 10.14.4.76:22
...
ci-info: | eth0 | True | 10.14.4.76 | 255.255.255.0 | 52:54:00:7b:ea:23 |
...
DHCPACK of 10.14.4.76 from 10.14.4.1
bound to 10.14.4.76 -- renewal in 17171 seconds.
However, when juju comes to tell mongo what its public facing address is, we pick 10.14.4.59 which is bogus but first in the list.
Jul 11 10:24:15 node0vm0 mongod.
Jul 11 10:24:15 node0vm0 mongod.
I don't think we have a reasonable course of action if the cloud is giving us bogus information here. There's no sane way of retrying this operation, and testing each address beforehand just in case the api is lying about an address is also not robust - not all addresses are valid in all contexts, but we do want to keep them around.
| Changed in juju-core: | |
| milestone: | 1.21-alpha1 → none |
| Patrick Hetu (patrick-hetu) wrote : | #9 |
I think I've got this bug using the local provider.
If got an external (eth0) network in 10.36.0.0/24 and an internal (lxcbr0) in 10.3.0.0/24 and got this errror:
[initandlisten] connection accepted from 10.36.0.35:42103 #2 (1 connection now open)
[conn2] assertion 16550 not authorized for query on local.system.
[conn2] ntoskip:0 ntoreturn:-1
[conn2] end connection 10.36.0.35:42103 (0 connections now open)
It looks like a juju is picking the 10.36.0.0/24 network instead of 10.3.0.0 like you said.
I was wondering if there is anything I can do to fix that?
| Dimiter Naydenov (dimitern) wrote : | #10 |
It looks like you might be hitting the same issue as in bug 1416928. Can you retry with the latest 1.21.3 proposed release and see if that resolves the issue? https:/


Using a loopback mount of the VM's image, I can see the bootstrap errors in cloud-init- output. log but there are no logs from mongo (/var/log/mongo doesn't exist)