Hello everyone, and sorry about the noise before. Here's a test plan that I think should satisfy all the requirements. First, we create a small ceph cluster. I used juju to add some machines with the following commands: 'juju add-machine --series=jammy' Afterwards, we ssh into the target machines and added the -proposed archives: $ cat /etc/apt/sources.list.d/ubuntu-jammy-proposed.list deb http://archive.ubuntu.com/ubuntu jammy-proposed main multiverse restricted universe We can verify that the 17.2.6 is going to be installed by running: $ apt-cache policy ceph ceph: Installed: 17.2.6-0ubuntu0.22.04.1 Candidate: 17.2.6-0ubuntu0.22.04.1 Version table: *** 17.2.6-0ubuntu0.22.04.1 500 500 http://archive.ubuntu.com/ubuntu jammy-proposed/main amd64 Packages 100 /var/lib/dpkg/status 17.2.5-0ubuntu0.22.04.3 500 500 http://nova.clouds.archive.ubuntu.com/ubuntu jammy-updates/main amd64 Packages 500 http://security.ubuntu.com/ubuntu jammy-security/main amd64 Packages 17.1.0-0ubuntu3 500 500 http://nova.clouds.archive.ubuntu.com/ubuntu jammy/main amd64 Packages With that in place, we deploy a small ceph cluster. I used 3 mons and 3 osd's, but anything should work. Once the cluster has been deployed we once again ssh into one of the target machines (in this case, one of the mons). As a precaution, we can test that ceph is running the proposed package: $ ceph-mon -v ceph version 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy (stable) Now, we ensure that rook-mgr isn't installed or running: $ sudo ceph mgr module ls MODULE balancer on (always on) crash on (always on) devicehealth on (always on) orchestrator on (always on) pg_autoscaler on (always on) progress on (always on) rbd_support on (always on) status on (always on) telemetry on (always on) volumes on (always on) iostat on nfs on restful on alerts - influx - insights - localpool - mirroring - osd_perf_query - osd_support - prometheus - selftest - snap_schedule - stats - telegraf - test_orchestrator - zabbix - Then, we install the rook-mgr module by hand: $ sudo apt install ceph-mgr-rook Next, we enable the module: $ sudo ceph mgr module enable rook We then check that the ceph cluster is in a healthy state and no modules have crashed by running: ubuntu@juju-9f12b1-ceph-0:~$ sudo ceph -s cluster: id: 026e3f56-1f5e-11ee-bf28-95ba6942eafd health: HEALTH_OK services: mon: 3 daemons, quorum juju-9f12b1-ceph-1,juju-9f12b1-ceph-2,juju-9f12b1-ceph-0 (age 9m) mgr: juju-9f12b1-ceph-2(active, since 6m), standbys: juju-9f12b1-ceph-1, juju-9f12b1-ceph-0 osd: 3 osds: 3 up, 3 in data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail And finally, we check that the rook module is up and running: $ sudo ceph mgr module ls MODULE balancer on (always on) crash on (always on) devicehealth on (always on) orchestrator on (always on) pg_autoscaler on (always on) progress on (always on) rbd_support on (always on) status on (always on) telemetry on (always on) volumes on (always on) iostat on nfs on restful on rook on alerts - influx - insights - localpool - mirroring - osd_perf_query - osd_support - prometheus - selftest - snap_schedule - stats - telegraf - test_orchestrator - zabbix - The process can be repeated with Kinetic instead, with identical results. Again, I apologize for not providing a correct test plan before. Please let me know if any further verification is needed.