2017-12-13 16:35:43 DEBUG juju.worker.peergrouper publish.go:43 API host ports have not changed
2017-12-13 16:35:43 DEBUG juju.worker.peergrouper desired.go:38 calculating desired peer group
2017-12-13 16:35:43 DEBUG juju.worker.peergrouper desired.go:43 members: ...
&peergrouper.machine{id: "0", wantsVote: true, hostPorts: [10.245.208.204:37017 127.0.0.1:37017 [::1]:37017]}: rs_id=1, rs_addr=10.245.208.204:37017
2017-12-13 16:35:43 DEBUG juju.worker.peergrouper desired.go:44 extra: []replicaset.Member(nil)
2017-12-13 16:35:43 DEBUG juju.worker.peergrouper desired.go:45 maxId: 1
2017-12-13 16:35:43 DEBUG juju.worker.peergrouper desired.go:117 assessing possible peer group changes:
2017-12-13 16:35:43 DEBUG juju.worker.peergrouper desired.go:124 machine "0" is already voting
2017-12-13 16:35:43 DEBUG juju.worker.peergrouper desired.go:142 assessed
2017-12-13 16:35:43 DEBUG juju.mongo mongo.go:288 selecting mongo peer hostPort by scope from [10.245.208.204:37017 127.0.0.1:37017 [::1]:37017]
2017-12-13 16:35:43 DEBUG juju.network address.go:370 selected "10.245.208.204:37017" as controller host:port, using scope selection
2017-12-13 16:35:43 DEBUG juju.mongo mongo.go:288 selecting mongo peer hostPort by scope from [10.245.208.204:37017 127.0.0.1:37017 [::1]:37017]
2017-12-13 16:35:43 DEBUG juju.network address.go:370 selected "10.245.208.204:37017" as controller host:port, using scope selection
2017-12-13 16:35:43 DEBUG juju.worker.peergrouper worker.go:494 no change in desired peer group, voting:
0: true
At this point, that is only a couple of minutes old.
It does claim that it wasn't able to find a space for the controllers, but it also seems to want to put 0, 1, 2 into a replicaset:
> db.controllers.find({"_id": "e"}).pretty()
{
"_id" : "e",
"cloud" : "foundations-maas", "model-uuid" : "f27da689-56d0-4c2f-8e80-4485b7e9d074", "machineids" : [
"0"
], "votingmachineids" : [ "0", "1",
"2"
], "mongo-space-name" : "", "mongo-space-state" : "invalid",
"txn-revno" : NumberLong(5),
"txn-queue" : [ ]
}
So far, nothing obvious. They all look to be up and happy.
$ juju status -m controller
...
Machine State DNS Inst id Series AZ Message
0 started 10.245.208.204 wccabc xenial default Deployed
1 started 10.245.208.200 fk44nr xenial zone2 Deployed
2 started 10.245.208.201 xcnwhm xenial zone3 Deployed
All are flagged green, so should be happy.
I have to figure out why the peergrouper isn't thinking they should be part of the group.
I don't see any errors happening, it just doesn't seem to think they are up and happy.
On a machine demonstrating this, we get: "2017-12- 13T16:35: 05.020Z" ),
"heartbeatInte rvalMillis" : NumberLong(2000),
"_id" : 1,
"name" : "10.245. 208.204: 37017",
"health" : 1,
"state" : 1,
"stateStr" : "PRIMARY",
"uptime" : 76881,
"optime" : {
"ts" : Timestamp( 1513182904, 5),
"t" : NumberLong(1)
},
"optimeDate" : ISODate( "2017-12- 13T16:35: 04Z"),
"electionTim e" : Timestamp( 1513106027, 1),
"electionDat e" : ISODate( "2017-12- 12T19:13: 47Z"),
"configVersi on" : 1,
"self" : true
rs.status()
{
"set" : "juju",
"date" : ISODate(
"myState" : 1,
"term" : NumberLong(1),
"members" : [
{
}
],
"ok" : 1
}
and
"protocolVersi on" : NumberLong(1),
"_id" : 1,
"host" : "10.245. 208.204: 37017",
"arbiterOnly " : false,
"buildIndexe s" : true,
"hidden" : false,
"priority" : 1,
"tags" : {
"juju- machine- id" : "0"
},
"slaveDelay" : NumberLong(0),
"votes" : 1
"chainingAllo wed" : true,
"heartbeatInt ervalMillis" : 2000,
"heartbeatTim eoutSecs" : 10,
"electionTime outMillis" : 10000,
"getLastError Modes" : {
> rs.config()
{
"_id" : "juju",
"version" : 1,
"members" : [
{
}
],
"settings" : {
},
"getLastError Defaults" : {
"w" : 1,
"wtimeout" : 0
"replicaSetId " : ObjectId( "5a302a6ac77b4c a613f5d47d" )
},
}
}
and machine-0.log is saying:
2017-12-13 16:35:43 DEBUG juju.worker. peergrouper publish.go:43 API host ports have not changed peergrouper desired.go:38 calculating desired peer group peergrouper desired.go:43 members: ... .machine{ id: "0", wantsVote: true, hostPorts: [10.245. 208.204: 37017 127.0.0.1:37017 [::1]:37017]}: rs_id=1, rs_addr= 10.245. 208.204: 37017 peergrouper desired.go:44 extra: []replicaset. Member( nil) peergrouper desired.go:45 maxId: 1 peergrouper desired.go:117 assessing possible peer group changes: peergrouper desired.go:124 machine "0" is already voting peergrouper desired.go:142 assessed 208.204: 37017 127.0.0.1:37017 [::1]:37017] 208.204: 37017" as controller host:port, using scope selection 208.204: 37017 127.0.0.1:37017 [::1]:37017] 208.204: 37017" as controller host:port, using scope selection peergrouper worker.go:494 no change in desired peer group, voting:
2017-12-13 16:35:43 DEBUG juju.worker.
2017-12-13 16:35:43 DEBUG juju.worker.
&peergrouper
2017-12-13 16:35:43 DEBUG juju.worker.
2017-12-13 16:35:43 DEBUG juju.worker.
2017-12-13 16:35:43 DEBUG juju.worker.
2017-12-13 16:35:43 DEBUG juju.worker.
2017-12-13 16:35:43 DEBUG juju.worker.
2017-12-13 16:35:43 DEBUG juju.mongo mongo.go:288 selecting mongo peer hostPort by scope from [10.245.
2017-12-13 16:35:43 DEBUG juju.network address.go:370 selected "10.245.
2017-12-13 16:35:43 DEBUG juju.mongo mongo.go:288 selecting mongo peer hostPort by scope from [10.245.
2017-12-13 16:35:43 DEBUG juju.network address.go:370 selected "10.245.
2017-12-13 16:35:43 DEBUG juju.worker.
0: true
At this point, that is only a couple of minutes old.
It does claim that it wasn't able to find a space for the controllers, but it also seems to want to put 0, 1, 2 into a replicaset: find({" _id": "e"}).pretty()
"model- uuid" : "f27da689- 56d0-4c2f- 8e80-4485b7e9d0 74",
"machineids" : [
"votingmachine ids" : [
"0",
"1",
"mongo- space-name" : "",
"mongo- space-state" : "invalid",
> db.controllers.
{
"_id" : "e",
"cloud" : "foundations-maas",
"0"
],
"2"
],
"txn-revno" : NumberLong(5),
"txn-queue" : [ ]
}
$ juju show-controller machines:
...
controller-
"0":
instance-id: wccabc
ha-status: ha-enabled
"1":
instance-id: fk44nr
ha-status: ha-pending
"2":
instance-id: xcnwhm
ha-status: ha-pending
...
So far, nothing obvious. They all look to be up and happy.
$ juju status -m controller
...
Machine State DNS Inst id Series AZ Message
0 started 10.245.208.204 wccabc xenial default Deployed
1 started 10.245.208.200 fk44nr xenial zone2 Deployed
2 started 10.245.208.201 xcnwhm xenial zone3 Deployed
All are flagged green, so should be happy.
I have to figure out why the peergrouper isn't thinking they should be part of the group.
I don't see any errors happening, it just doesn't seem to think they are up and happy.
still needs some digging.