Comment 14 for bug 2003530

Revision history for this message
Luciano Lo Giudice (lmlogiudice) wrote :

Hello everyone, and sorry about the noise before. Here's a test plan that I think should satisfy all the requirements.

First, we create a small ceph cluster. I used juju to add some machines with the following commands:

'juju add-machine --series=jammy'

Afterwards, we ssh into the target machines and added the -proposed archives:

$ cat /etc/apt/sources.list.d/ubuntu-jammy-proposed.list
deb http://archive.ubuntu.com/ubuntu jammy-proposed main multiverse restricted universe

We can verify that the 17.2.6 is going to be installed by running:

$ apt-cache policy ceph
ceph:
  Installed: 17.2.6-0ubuntu0.22.04.1
  Candidate: 17.2.6-0ubuntu0.22.04.1
  Version table:
 *** 17.2.6-0ubuntu0.22.04.1 500
        500 http://archive.ubuntu.com/ubuntu jammy-proposed/main amd64 Packages
        100 /var/lib/dpkg/status
     17.2.5-0ubuntu0.22.04.3 500
        500 http://nova.clouds.archive.ubuntu.com/ubuntu jammy-updates/main amd64 Packages
        500 http://security.ubuntu.com/ubuntu jammy-security/main amd64 Packages
     17.1.0-0ubuntu3 500
        500 http://nova.clouds.archive.ubuntu.com/ubuntu jammy/main amd64 Packages

With that in place, we deploy a small ceph cluster. I used 3 mons and 3 osd's, but anything should work.

Once the cluster has been deployed we once again ssh into one of the target machines (in this case, one of the mons).

As a precaution, we can test that ceph is running the proposed package:

$ ceph-mon -v
ceph version 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy (stable)

Now, we ensure that rook-mgr isn't installed or running:

$ sudo ceph mgr module ls

MODULE
balancer on (always on)
crash on (always on)
devicehealth on (always on)
orchestrator on (always on)
pg_autoscaler on (always on)
progress on (always on)
rbd_support on (always on)
status on (always on)
telemetry on (always on)
volumes on (always on)
iostat on
nfs on
restful on
alerts -
influx -
insights -
localpool -
mirroring -
osd_perf_query -
osd_support -
prometheus -
selftest -
snap_schedule -
stats -
telegraf -
test_orchestrator -
zabbix -

Then, we install the rook-mgr module by hand:

$ sudo apt install ceph-mgr-rook

Next, we enable the module:

$ sudo ceph mgr module enable rook

We then check that the ceph cluster is in a healthy state and no modules have crashed by running:

ubuntu@juju-9f12b1-ceph-0:~$ sudo ceph -s
  cluster:
    id: 026e3f56-1f5e-11ee-bf28-95ba6942eafd
    health: HEALTH_OK

  services:
    mon: 3 daemons, quorum juju-9f12b1-ceph-1,juju-9f12b1-ceph-2,juju-9f12b1-ceph-0 (age 9m)
    mgr: juju-9f12b1-ceph-2(active, since 6m), standbys: juju-9f12b1-ceph-1, juju-9f12b1-ceph-0
    osd: 3 osds: 3 up, 3 in

  data:
    pools: 0 pools, 0 pgs
    objects: 0 objects, 0 B
    usage: 0 B used, 0 B / 0 B avail

And finally, we check that the rook module is up and running:

$ sudo ceph mgr module ls

MODULE
balancer on (always on)
crash on (always on)
devicehealth on (always on)
orchestrator on (always on)
pg_autoscaler on (always on)
progress on (always on)
rbd_support on (always on)
status on (always on)
telemetry on (always on)
volumes on (always on)
iostat on
nfs on
restful on
rook on
alerts -
influx -
insights -
localpool -
mirroring -
osd_perf_query -
osd_support -
prometheus -
selftest -
snap_schedule -
stats -
telegraf -
test_orchestrator -
zabbix -

The process can be repeated with Kinetic instead, with identical results.

Again, I apologize for not providing a correct test plan before. Please let me know if any further verification is needed.