2018-04-05 22:49:32 +00:00
|
|
|
Upgrading the MDS Cluster
|
|
|
|
=========================
|
|
|
|
|
|
|
|
Currently the MDS cluster does not have built-in versioning or file system
|
|
|
|
flags to support seamless upgrades of the MDSs without potentially causing
|
|
|
|
assertions or other faults due to incompatible messages or other functional
|
|
|
|
differences. For this reason, it's necessary during any cluster upgrade to
|
|
|
|
reduce the number of active MDS for a file system to one first so that two
|
|
|
|
active MDS do not communicate with different versions. Further, it's also
|
2018-09-18 03:19:18 +00:00
|
|
|
necessary to take standbys offline as any new CompatSet flags will propagate
|
2018-04-05 22:49:32 +00:00
|
|
|
via the MDSMap to all MDS and cause older MDS to suicide.
|
|
|
|
|
|
|
|
The proper sequence for upgrading the MDS cluster is:
|
|
|
|
|
|
|
|
1. Reduce the number of ranks to 1:
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
ceph fs set <fs_name> max_mds 1
|
|
|
|
|
2018-09-18 22:29:00 +00:00
|
|
|
2. Wait for cluster to stop non-zero ranks where only rank 0 is active and the rest are standbys.
|
2018-04-05 22:49:32 +00:00
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
ceph status # wait for MDS to finish stopping
|
|
|
|
|
|
|
|
3. Take all standbys offline, e.g. using systemctl:
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
systemctl stop ceph-mds.target
|
|
|
|
|
2018-04-13 23:48:36 +00:00
|
|
|
4. Confirm only one MDS is online and is rank 0 for your FS:
|
2018-04-05 22:49:32 +00:00
|
|
|
|
|
|
|
::
|
|
|
|
|
2018-04-13 23:48:36 +00:00
|
|
|
ceph status
|
|
|
|
|
|
|
|
5. Upgrade the single active MDS, e.g. using systemctl:
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
# use package manager to update cluster
|
2018-04-05 22:49:32 +00:00
|
|
|
systemctl restart ceph-mds.target
|
|
|
|
|
2018-04-13 23:48:36 +00:00
|
|
|
6. Upgrade/start the standby daemons.
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
# use package manager to update cluster
|
|
|
|
systemctl restart ceph-mds.target
|
2018-04-05 22:49:32 +00:00
|
|
|
|
2018-04-13 23:48:36 +00:00
|
|
|
7. Restore the previous max_mds for your cluster:
|
2018-04-05 22:49:32 +00:00
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
ceph fs set <fs_name> max_mds <old_max_mds>
|
|
|
|
|
2016-02-29 14:19:50 +00:00
|
|
|
|
2019-09-09 19:36:04 +00:00
|
|
|
Upgrading pre-Firefly file systems past Jewel
|
|
|
|
=============================================
|
2016-02-29 14:19:50 +00:00
|
|
|
|
|
|
|
.. tip::
|
|
|
|
|
2019-09-09 19:36:04 +00:00
|
|
|
This advice only applies to users with file systems
|
2016-02-29 14:19:50 +00:00
|
|
|
created using versions of Ceph older than *Firefly* (0.80).
|
2019-09-09 19:36:04 +00:00
|
|
|
Users creating new file systems may disregard this advice.
|
2016-02-29 14:19:50 +00:00
|
|
|
|
|
|
|
Pre-firefly versions of Ceph used a now-deprecated format
|
|
|
|
for storing CephFS directory objects, called TMAPs. Support
|
|
|
|
for reading these in RADOS will be removed after the Jewel
|
|
|
|
release of Ceph, so for upgrading CephFS users it is important
|
|
|
|
to ensure that any old directory objects have been converted.
|
|
|
|
|
|
|
|
After installing Jewel on all your MDS and OSD servers, and restarting
|
|
|
|
the services, run the following command:
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
cephfs-data-scan tmap_upgrade <metadata pool name>
|
|
|
|
|
|
|
|
This only needs to be run once, and it is not necessary to
|
|
|
|
stop any other services while it runs. The command may take some
|
|
|
|
time to execute, as it iterates overall objects in your metadata
|
2019-09-09 19:36:04 +00:00
|
|
|
pool. It is safe to continue using your file system as normal while
|
2016-02-29 14:19:50 +00:00
|
|
|
it executes. If the command aborts for any reason, it is safe
|
|
|
|
to simply run it again.
|
|
|
|
|
2019-09-09 19:36:04 +00:00
|
|
|
If you are upgrading a pre-Firefly CephFS file system to a newer Ceph version
|
2016-02-29 14:19:50 +00:00
|
|
|
than Jewel, you must first upgrade to Jewel and run the ``tmap_upgrade``
|
|
|
|
command before completing your upgrade to the latest version.
|
|
|
|
|