doc/releases/nautilus: more detail around ceph-volume step

Recommend host reboots to verify ceph-volume adoption worked.

Signed-off-by: Sage Weil <sage@redhat.com>
This commit is contained in:
Sage Weil 2019-03-14 14:43:08 -05:00
parent f80051b4ce
commit 52c447a49a

View File

@ -216,13 +216,31 @@ Instructions
"ceph version 14.2.0 (...) nautilus (stable)": 22,
}
#. Scan for any OSDs deployed with the old ceph-disk tool to ensure
that ceph-volume can activate them after a host reboot. On each
host containing OSDs,::
#. If there are any OSDs in the cluster deployed with ceph-disk (e.g.,
almost any OSDs that were created before the Mimic release), you
need to tell ceph-volume to adopt responsibility for starting the
daemons. On each host containing OSDs, ensure the OSDs are
currently running, and then::
# ceph-volume simple scan
# ceph-volume simple activate --all
We recommend that each OSD host be rebooted following this step to
verify that the OSDs start up automatically.
Note that ceph-volume doesn't have the same hot-plug capability
that ceph-disk did, where a newly attached disk is automatically
detected via udev events. If the OSD isn't currently running when the
above ``scan`` command is run, or a ceph-disk-based OSD is moved to
a new host, or the host OSD is reinstalled, or the
``/etc/ceph/osd`` directory is lost, you will need to scan the main
data partition for each ceph-disk OSD explicitly. For example,::
# ceph-volume simple scan /dev/sdb1
The output will include the appopriate ``ceph-volume simple
activate`` command to enable the OSD.
#. Upgrade all CephFS MDS daemons. For each CephFS file system,
#. Reduce the number of ranks to 1. (Make note of the original