remove-unit doesn't take OSDs down

Bug #1629679 reported by James Troup
44
This bug affects 9 people
Affects Status Importance Assigned to Milestone
Ceph OSD Charm
Triaged
Wishlist
Unassigned
OpenStack Ceph Charm (Retired)
Won't Fix
Medium
Unassigned
ceph (Juju Charms Collection)
Invalid
Medium
Unassigned
ceph-osd (Juju Charms Collection)
Invalid
Medium
Unassigned

Bug Description

I ran 'juju destroy ceph-osd/0' and expected this to take the OSDs on
ceph-osd/0 down and out. It didn't; even when ceph-osd/0 was
completely gone, the OSDs were still up and running.

Given ceph is currently run almost exclusively on bare metal (and
therefor entirely possibly not the last unit on the machine), I think
it would make sense for the charms to not assume the machine hosting
the unit is going away and instead explicitly take down the OSDs. And
perhaps stop them from coming back up again to avoid epic confusion on
reboot?

James Troup (elmo)
tags: added: canonical-is
tags: added: canonical-bootstack
description: updated
Revision history for this message
James Troup (elmo) wrote :

FAOD this is not unit specific; I just did 'juju destroy-service ceph-osd' and still have a functional ceph cluster.

Revision history for this message
James Troup (elmo) wrote :

Additionally, the "tuning" done by newer versions of the charm persists after the unit or service has been destroyed. This seems extremely undesirable.

Revision history for this message
James Troup (elmo) wrote :

And the alternatives handling for ceph.conf...

Revision history for this message
Chris MacNaughton (chris.macnaughton) wrote : Re: [Bug 1629679] Re: destroy-unit doesn't take OSDs down

The ceph charms have no stop hook, meaning that there's no expectation of it removing anything on removal, unfortunately

> On Oct 2, 2016, at 19:25, James Troup <email address hidden> wrote:
>
> And the alternatives handling for ceph.conf...
>
> --
> You received this bug notification because you are a member of OpenStack
> Charmers, which is subscribed to ceph in Juju Charms Collection.
> https://bugs.launchpad.net/bugs/1629679
>
> Title:
> destroy-unit doesn't take OSDs down
>
> Status in ceph package in Juju Charms Collection:
> New
>
> Bug description:
> I ran 'juju destroy ceph-osd/0' and expected this to take the OSDs on
> ceph-osd/0 down and out. It didn't; even when ceph-osd/0 was
> completely gone, the OSDs were still up and running.
>
> Given ceph is currently run almost exclusively on bare metal (and
> therefor entirely possibly not the last unit on the machine), I think
> it would make sense for the charms to not assume the machine hosting
> the unit is going away and instead explicitly take down the OSDs. And
> perhaps stop them from coming back up again to avoid epic confusion on
> reboot?
>
> To manage notifications about this bug go to:
> https://bugs.launchpad.net/charms/+source/ceph/+bug/1629679/+subscriptions

Revision history for this message
Sandor Zeestraten (szeestraten) wrote : Re: destroy-unit doesn't take OSDs down

Is this something that is in the plans for future releases?
If not, could we at least state the expected behaviour more explicitly in the readme of the charms?

Revision history for this message
James Page (james-page) wrote :

I'll raise this at the next charms IRC meeting for discussion - I think we should teardown and running services when a unit is destroyed (via its stop hook), ensuring that the machine is clean; we can leave the OSD state on disk (which can be overrwritten during re-deployment if need be).

Changed in ceph (Juju Charms Collection):
status: New → Triaged
importance: Undecided → Medium
Changed in ceph-osd (Juju Charms Collection):
status: New → Triaged
importance: Undecided → Medium
James Page (james-page)
Changed in charm-ceph:
importance: Undecided → Medium
status: New → Triaged
Changed in ceph (Juju Charms Collection):
status: Triaged → Invalid
Changed in charm-ceph-osd:
importance: Undecided → Medium
status: New → Triaged
Changed in ceph-osd (Juju Charms Collection):
status: Triaged → Invalid
Changed in charm-ceph:
status: Triaged → Won't Fix
tags: added: scaleback
Revision history for this message
Ryan Beisner (1chb1n) wrote :

Circling back to this topic, let's be mindful of our position about not automatically wiping block devices.

Changed in charm-ceph-osd:
status: In Progress → New
Changed in charm-ceph-osd:
status: New → Triaged
importance: Medium → Wishlist
Revision history for this message
Peter Matulis (petermatulis) wrote :

Changed the title as `destroy-unit` is now `remove-unit`.

summary: - destroy-unit doesn't take OSDs down
+ remove-unit doesn't take OSDs down
Revision history for this message
OpenStack Infra (hudson-openstack) wrote : Change abandoned on charm-ceph-osd (master)

Change abandoned by "James Page <email address hidden>" on branch: master
Review: https://review.opendev.org/c/openstack/charm-ceph-osd/+/659536
Reason: This review is > 12 weeks without comment, and failed testing the last time it was checked. We are abandoning this for now. Feel free to reactivate the review by pressing the restore button and leaving a 'recheck' comment to get fresh test results.

To post a comment you must log in.
This report contains Public information  
Everyone can see this information.

Duplicates of this bug

Other bug subscribers

Remote bug watches

Bug watches keep track of this bug in other bug trackers.