Hello everyone, and sorry about the noise before. Here's a test plan that I think should satisfy all the requirements.
First, we create a small ceph cluster. I used juju to add some machines with the following commands:
'juju add-machine --series=jammy'
Afterwards, we ssh into the target machines and added the -proposed archives:
$ cat /etc/apt/sources.list.d/ubuntu-jammy-proposed.list deb http://archive.ubuntu.com/ubuntu jammy-proposed main multiverse restricted universe
We can verify that the 17.2.6 is going to be installed by running:
$ apt-cache policy ceph ceph: Installed: 17.2.6-0ubuntu0.22.04.1 Candidate: 17.2.6-0ubuntu0.22.04.1 Version table: *** 17.2.6-0ubuntu0.22.04.1 500 500 http://archive.ubuntu.com/ubuntu jammy-proposed/main amd64 Packages 100 /var/lib/dpkg/status 17.2.5-0ubuntu0.22.04.3 500 500 http://nova.clouds.archive.ubuntu.com/ubuntu jammy-updates/main amd64 Packages 500 http://security.ubuntu.com/ubuntu jammy-security/main amd64 Packages 17.1.0-0ubuntu3 500 500 http://nova.clouds.archive.ubuntu.com/ubuntu jammy/main amd64 Packages
With that in place, we deploy a small ceph cluster. I used 3 mons and 3 osd's, but anything should work.
Once the cluster has been deployed we once again ssh into one of the target machines (in this case, one of the mons).
As a precaution, we can test that ceph is running the proposed package:
$ ceph-mon -v ceph version 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy (stable)
Now, we ensure that rook-mgr isn't installed or running:
$ sudo ceph mgr module ls
MODULE balancer on (always on) crash on (always on) devicehealth on (always on) orchestrator on (always on) pg_autoscaler on (always on) progress on (always on) rbd_support on (always on) status on (always on) telemetry on (always on) volumes on (always on) iostat on nfs on restful on alerts - influx - insights - localpool - mirroring - osd_perf_query - osd_support - prometheus - selftest - snap_schedule - stats - telegraf - test_orchestrator - zabbix -
Then, we install the rook-mgr module by hand:
$ sudo apt install ceph-mgr-rook
Next, we enable the module:
$ sudo ceph mgr module enable rook
We then check that the ceph cluster is in a healthy state and no modules have crashed by running:
ubuntu@juju-9f12b1-ceph-0:~$ sudo ceph -s cluster: id: 026e3f56-1f5e-11ee-bf28-95ba6942eafd health: HEALTH_OK
services: mon: 3 daemons, quorum juju-9f12b1-ceph-1,juju-9f12b1-ceph-2,juju-9f12b1-ceph-0 (age 9m) mgr: juju-9f12b1-ceph-2(active, since 6m), standbys: juju-9f12b1-ceph-1, juju-9f12b1-ceph-0 osd: 3 osds: 3 up, 3 in
data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail
And finally, we check that the rook module is up and running:
MODULE balancer on (always on) crash on (always on) devicehealth on (always on) orchestrator on (always on) pg_autoscaler on (always on) progress on (always on) rbd_support on (always on) status on (always on) telemetry on (always on) volumes on (always on) iostat on nfs on restful on rook on alerts - influx - insights - localpool - mirroring - osd_perf_query - osd_support - prometheus - selftest - snap_schedule - stats - telegraf - test_orchestrator - zabbix -
The process can be repeated with Kinetic instead, with identical results.
Again, I apologize for not providing a correct test plan before. Please let me know if any further verification is needed.
Hello everyone, and sorry about the noise before. Here's a test plan that I think should satisfy all the requirements.
First, we create a small ceph cluster. I used juju to add some machines with the following commands:
'juju add-machine --series=jammy'
Afterwards, we ssh into the target machines and added the -proposed archives:
$ cat /etc/apt/ sources. list.d/ ubuntu- jammy-proposed. list archive. ubuntu. com/ubuntu jammy-proposed main multiverse restricted universe
deb http://
We can verify that the 17.2.6 is going to be installed by running:
$ apt-cache policy ceph 0ubuntu0. 22.04.1 0ubuntu0. 22.04.1 0ubuntu0. 22.04.1 500 archive. ubuntu. com/ubuntu jammy-proposed/main amd64 Packages dpkg/status 2.5-0ubuntu0. 22.04.3 500 nova.clouds. archive. ubuntu. com/ubuntu jammy-updates/main amd64 Packages security. ubuntu. com/ubuntu jammy-security/main amd64 Packages 1.0-0ubuntu3 500 nova.clouds. archive. ubuntu. com/ubuntu jammy/main amd64 Packages
ceph:
Installed: 17.2.6-
Candidate: 17.2.6-
Version table:
*** 17.2.6-
500 http://
100 /var/lib/
17.
500 http://
500 http://
17.
500 http://
With that in place, we deploy a small ceph cluster. I used 3 mons and 3 osd's, but anything should work.
Once the cluster has been deployed we once again ssh into one of the target machines (in this case, one of the mons).
As a precaution, we can test that ceph is running the proposed package:
$ ceph-mon -v 80e08f1ab989c7c df3064446a5) quincy (stable)
ceph version 17.2.6 (d7ff0d10654d22
Now, we ensure that rook-mgr isn't installed or running:
$ sudo ceph mgr module ls
MODULE
balancer on (always on)
crash on (always on)
devicehealth on (always on)
orchestrator on (always on)
pg_autoscaler on (always on)
progress on (always on)
rbd_support on (always on)
status on (always on)
telemetry on (always on)
volumes on (always on)
iostat on
nfs on
restful on
alerts -
influx -
insights -
localpool -
mirroring -
osd_perf_query -
osd_support -
prometheus -
selftest -
snap_schedule -
stats -
telegraf -
test_orchestrator -
zabbix -
Then, we install the rook-mgr module by hand:
$ sudo apt install ceph-mgr-rook
Next, we enable the module:
$ sudo ceph mgr module enable rook
We then check that the ceph cluster is in a healthy state and no modules have crashed by running:
ubuntu@ juju-9f12b1- ceph-0: ~$ sudo ceph -s 1f5e-11ee- bf28-95ba6942ea fd
cluster:
id: 026e3f56-
health: HEALTH_OK
services: ceph-1, juju-9f12b1- ceph-2, juju-9f12b1- ceph-0 (age 9m) ceph-2( active, since 6m), standbys: juju-9f12b1-ceph-1, juju-9f12b1-ceph-0
mon: 3 daemons, quorum juju-9f12b1-
mgr: juju-9f12b1-
osd: 3 osds: 3 up, 3 in
data:
pools: 0 pools, 0 pgs
objects: 0 objects, 0 B
usage: 0 B used, 0 B / 0 B avail
And finally, we check that the rook module is up and running:
$ sudo ceph mgr module ls
MODULE
balancer on (always on)
crash on (always on)
devicehealth on (always on)
orchestrator on (always on)
pg_autoscaler on (always on)
progress on (always on)
rbd_support on (always on)
status on (always on)
telemetry on (always on)
volumes on (always on)
iostat on
nfs on
restful on
rook on
alerts -
influx -
insights -
localpool -
mirroring -
osd_perf_query -
osd_support -
prometheus -
selftest -
snap_schedule -
stats -
telegraf -
test_orchestrator -
zabbix -
The process can be repeated with Kinetic instead, with identical results.
Again, I apologize for not providing a correct test plan before. Please let me know if any further verification is needed.