diff --git a/doc/releases/nautilus.rst b/doc/releases/nautilus.rst index c2892afb431..09009ac2a1b 100644 --- a/doc/releases/nautilus.rst +++ b/doc/releases/nautilus.rst @@ -216,13 +216,31 @@ Instructions "ceph version 14.2.0 (...) nautilus (stable)": 22, } -#. Scan for any OSDs deployed with the old ceph-disk tool to ensure - that ceph-volume can activate them after a host reboot. On each - host containing OSDs,:: +#. If there are any OSDs in the cluster deployed with ceph-disk (e.g., + almost any OSDs that were created before the Mimic release), you + need to tell ceph-volume to adopt responsibility for starting the + daemons. On each host containing OSDs, ensure the OSDs are + currently running, and then:: # ceph-volume simple scan # ceph-volume simple activate --all + We recommend that each OSD host be rebooted following this step to + verify that the OSDs start up automatically. + + Note that ceph-volume doesn't have the same hot-plug capability + that ceph-disk did, where a newly attached disk is automatically + detected via udev events. If the OSD isn't currently running when the + above ``scan`` command is run, or a ceph-disk-based OSD is moved to + a new host, or the host OSD is reinstalled, or the + ``/etc/ceph/osd`` directory is lost, you will need to scan the main + data partition for each ceph-disk OSD explicitly. For example,:: + + # ceph-volume simple scan /dev/sdb1 + + The output will include the appopriate ``ceph-volume simple + activate`` command to enable the OSD. + #. Upgrade all CephFS MDS daemons. For each CephFS file system, #. Reduce the number of ranks to 1. (Make note of the original