ceph/doc/man/8/ceph-deploy.rst
Jason Dillaman 65c8733b56 librados: revert librados3/libradoscc back to librados2
For backwards compatibility and upgrade reasons, the librados2
API needs to be preserved and it needs to continue to be compatible
with dependent libraries like librbd1.

Signed-off-by: Jason Dillaman <dillaman@redhat.com>
2019-02-20 14:59:26 -05:00

530 lines
18 KiB
ReStructuredText

:orphan:
.. _ceph-deploy:
=====================================
ceph-deploy -- Ceph deployment tool
=====================================
.. program:: ceph-deploy
Synopsis
========
| **ceph-deploy** **new** [*initial-monitor-node(s)*]
| **ceph-deploy** **install** [*ceph-node*] [*ceph-node*...]
| **ceph-deploy** **mon** *create-initial*
| **ceph-deploy** **osd** *create* *--data* *device* *ceph-node*
| **ceph-deploy** **admin** [*admin-node*][*ceph-node*...]
| **ceph-deploy** **purgedata** [*ceph-node*][*ceph-node*...]
| **ceph-deploy** **forgetkeys**
Description
===========
:program:`ceph-deploy` is a tool which allows easy and quick deployment of a
Ceph cluster without involving complex and detailed manual configuration. It
uses ssh to gain access to other Ceph nodes from the admin node, sudo for
administrator privileges on them and the underlying Python scripts automates
the manual process of Ceph installation on each node from the admin node itself.
It can be easily run on an workstation and doesn't require servers, databases or
any other automated tools. With :program:`ceph-deploy`, it is really easy to set
up and take down a cluster. However, it is not a generic deployment tool. It is
a specific tool which is designed for those who want to get Ceph up and running
quickly with only the unavoidable initial configuration settings and without the
overhead of installing other tools like ``Chef``, ``Puppet`` or ``Juju``. Those
who want to customize security settings, partitions or directory locations and
want to set up a cluster following detailed manual steps, should use other tools
i.e, ``Chef``, ``Puppet``, ``Juju`` or ``Crowbar``.
With :program:`ceph-deploy`, you can install Ceph packages on remote nodes,
create a cluster, add monitors, gather/forget keys, add OSDs and metadata
servers, configure admin hosts or take down the cluster.
Commands
========
new
---
Start deploying a new cluster and write a configuration file and keyring for it.
It tries to copy ssh keys from admin node to gain passwordless ssh to monitor
node(s), validates host IP, creates a cluster with a new initial monitor node or
nodes for monitor quorum, a ceph configuration file, a monitor secret keyring and
a log file for the new cluster. It populates the newly created Ceph configuration
file with ``fsid`` of cluster, hostnames and IP addresses of initial monitor
members under ``[global]`` section.
Usage::
ceph-deploy new [MON][MON...]
Here, [MON] is the initial monitor hostname (short hostname i.e, ``hostname -s``).
Other options like :option:`--no-ssh-copykey`, :option:`--fsid`,
:option:`--cluster-network` and :option:`--public-network` can also be used with
this command.
If more than one network interface is used, ``public network`` setting has to be
added under ``[global]`` section of Ceph configuration file. If the public subnet
is given, ``new`` command will choose the one IP from the remote host that exists
within the subnet range. Public network can also be added at runtime using
:option:`--public-network` option with the command as mentioned above.
install
-------
Install Ceph packages on remote hosts. As a first step it installs
``yum-plugin-priorities`` in admin and other nodes using passwordless ssh and sudo
so that Ceph packages from upstream repository get more priority. It then detects
the platform and distribution for the hosts and installs Ceph normally by
downloading distro compatible packages if adequate repo for Ceph is already added.
``--release`` flag is used to get the latest release for installation. During
detection of platform and distribution before installation, if it finds the
``distro.init`` to be ``sysvinit`` (Fedora, CentOS/RHEL etc), it doesn't allow
installation with custom cluster name and uses the default name ``ceph`` for the
cluster.
If the user explicitly specifies a custom repo url with :option:`--repo-url` for
installation, anything detected from the configuration will be overridden and
the custom repository location will be used for installation of Ceph packages.
If required, valid custom repositories are also detected and installed. In case
of installation from a custom repo a boolean is used to determine the logic
needed to proceed with a custom repo installation. A custom repo install helper
is used that goes through config checks to retrieve repos (and any extra repos
defined) and installs them. ``cd_conf`` is the object built from ``argparse``
that holds the flags and information needed to determine what metadata from the
configuration is to be used.
A user can also opt to install only the repository without installing Ceph and
its dependencies by using :option:`--repo` option.
Usage::
ceph-deploy install [HOST][HOST...]
Here, [HOST] is/are the host node(s) where Ceph is to be installed.
An option ``--release`` is used to install a release known as CODENAME
(default: firefly).
Other options like :option:`--testing`, :option:`--dev`, :option:`--adjust-repos`,
:option:`--no-adjust-repos`, :option:`--repo`, :option:`--local-mirror`,
:option:`--repo-url` and :option:`--gpg-url` can also be used with this command.
mds
---
Deploy Ceph mds on remote hosts. A metadata server is needed to use CephFS and
the ``mds`` command is used to create one on the desired host node. It uses the
subcommand ``create`` to do so. ``create`` first gets the hostname and distro
information of the desired mds host. It then tries to read the ``bootstrap-mds``
key for the cluster and deploy it in the desired host. The key generally has a
format of ``{cluster}.bootstrap-mds.keyring``. If it doesn't finds a keyring,
it runs ``gatherkeys`` to get the keyring. It then creates a mds on the desired
host under the path ``/var/lib/ceph/mds/`` in ``/var/lib/ceph/mds/{cluster}-{name}``
format and a bootstrap keyring under ``/var/lib/ceph/bootstrap-mds/`` in
``/var/lib/ceph/bootstrap-mds/{cluster}.keyring`` format. It then runs appropriate
commands based on ``distro.init`` to start the ``mds``.
Usage::
ceph-deploy mds create [HOST[:DAEMON-NAME]] [HOST[:DAEMON-NAME]...]
The [DAEMON-NAME] is optional.
mon
---
Deploy Ceph monitor on remote hosts. ``mon`` makes use of certain subcommands
to deploy Ceph monitors on other nodes.
Subcommand ``create-initial`` deploys for monitors defined in
``mon initial members`` under ``[global]`` section in Ceph configuration file,
wait until they form quorum and then gatherkeys, reporting the monitor status
along the process. If monitors don't form quorum the command will eventually
time out.
Usage::
ceph-deploy mon create-initial
Subcommand ``create`` is used to deploy Ceph monitors by explicitly specifying
the hosts which are desired to be made monitors. If no hosts are specified it
will default to use the ``mon initial members`` defined under ``[global]``
section of Ceph configuration file. ``create`` first detects platform and distro
for desired hosts and checks if hostname is compatible for deployment. It then
uses the monitor keyring initially created using ``new`` command and deploys the
monitor in desired host. If multiple hosts were specified during ``new`` command
i.e, if there are multiple hosts in ``mon initial members`` and multiple keyrings
were created then a concatenated keyring is used for deployment of monitors. In
this process a keyring parser is used which looks for ``[entity]`` sections in
monitor keyrings and returns a list of those sections. A helper is then used to
collect all keyrings into a single blob that will be used to inject it to monitors
with :option:`--mkfs` on remote nodes. All keyring files are concatenated to be
in a directory ending with ``.keyring``. During this process the helper uses list
of sections returned by keyring parser to check if an entity is already present
in a keyring and if not, adds it. The concatenated keyring is used for deployment
of monitors to desired multiple hosts.
Usage::
ceph-deploy mon create [HOST] [HOST...]
Here, [HOST] is hostname of desired monitor host(s).
Subcommand ``add`` is used to add a monitor to an existing cluster. It first
detects platform and distro for desired host and checks if hostname is compatible
for deployment. It then uses the monitor keyring, ensures configuration for new
monitor host and adds the monitor to the cluster. If the section for the monitor
exists and defines a monitor address that will be used, otherwise it will fallback by
resolving the hostname to an IP. If :option:`--address` is used it will override
all other options. After adding the monitor to the cluster, it gives it some time
to start. It then looks for any monitor errors and checks monitor status. Monitor
errors arise if the monitor is not added in ``mon initial members``, if it doesn't
exist in ``monmap`` and if neither ``public_addr`` nor ``public_network`` keys
were defined for monitors. Under such conditions, monitors may not be able to
form quorum. Monitor status tells if the monitor is up and running normally. The
status is checked by running ``ceph daemon mon.hostname mon_status`` on remote
end which provides the output and returns a boolean status of what is going on.
``False`` means a monitor that is not fine even if it is up and running, while
``True`` means the monitor is up and running correctly.
Usage::
ceph-deploy mon add [HOST]
ceph-deploy mon add [HOST] --address [IP]
Here, [HOST] is the hostname and [IP] is the IP address of the desired monitor
node. Please note, unlike other ``mon`` subcommands, only one node can be
specified at a time.
Subcommand ``destroy`` is used to completely remove monitors on remote hosts.
It takes hostnames as arguments. It stops the monitor, verifies if ``ceph-mon``
daemon really stopped, creates an archive directory ``mon-remove`` under
``/var/lib/ceph/``, archives old monitor directory in
``{cluster}-{hostname}-{stamp}`` format in it and removes the monitor from
cluster by running ``ceph remove...`` command.
Usage::
ceph-deploy mon destroy [HOST] [HOST...]
Here, [HOST] is hostname of monitor that is to be removed.
gatherkeys
----------
Gather authentication keys for provisioning new nodes. It takes hostnames as
arguments. It checks for and fetches ``client.admin`` keyring, monitor keyring
and ``bootstrap-mds/bootstrap-osd`` keyring from monitor host. These
authentication keys are used when new ``monitors/OSDs/MDS`` are added to the
cluster.
Usage::
ceph-deploy gatherkeys [HOST] [HOST...]
Here, [HOST] is hostname of the monitor from where keys are to be pulled.
disk
----
Manage disks on a remote host. It actually triggers the ``ceph-volume`` utility
and its subcommands to manage disks.
Subcommand ``list`` lists disk partitions and Ceph OSDs.
Usage::
ceph-deploy disk list HOST
Subcommand ``zap`` zaps/erases/destroys a device's partition table and
contents. It actually uses ``ceph-volume lvm zap`` remotely, alternatively
allowing someone to remove the Ceph metadata from the logical volume.
osd
---
Manage OSDs by preparing data disk on remote host. ``osd`` makes use of certain
subcommands for managing OSDs.
Subcommand ``create`` prepares a device for Ceph OSD. It first checks against
multiple OSDs getting created and warns about the possibility of more than the
recommended which would cause issues with max allowed PIDs in a system. It then
reads the bootstrap-osd key for the cluster or writes the bootstrap key if not
found.
It then uses :program:`ceph-volume` utility's ``lvm create`` subcommand to
prepare the disk, (and journal if using filestore) and deploy the OSD on the desired host.
Once prepared, it gives some time to the OSD to start and checks for any
possible errors and if found, reports to the user.
Bluestore Usage::
ceph-deploy osd create --data DISK HOST
Filestore Usage::
ceph-deploy osd create --data DISK --journal JOURNAL HOST
.. note:: For other flags available, please see the man page or the --help menu
on ceph-deploy osd create
Subcommand ``list`` lists devices associated to Ceph as part of an OSD.
It uses the ``ceph-volume lvm list`` output that has a rich output, mapping
OSDs to devices and other interesting information about the OSD setup.
Usage::
ceph-deploy osd list HOST
admin
-----
Push configuration and ``client.admin`` key to a remote host. It takes
the ``{cluster}.client.admin.keyring`` from admin node and writes it under
``/etc/ceph`` directory of desired node.
Usage::
ceph-deploy admin [HOST] [HOST...]
Here, [HOST] is desired host to be configured for Ceph administration.
config
------
Push/pull configuration file to/from a remote host. It uses ``push`` subcommand
to takes the configuration file from admin host and write it to remote host under
``/etc/ceph`` directory. It uses ``pull`` subcommand to do the opposite i.e, pull
the configuration file under ``/etc/ceph`` directory of remote host to admin node.
Usage::
ceph-deploy config push [HOST] [HOST...]
ceph-deploy config pull [HOST] [HOST...]
Here, [HOST] is the hostname of the node where config file will be pushed to or
pulled from.
uninstall
---------
Remove Ceph packages from remote hosts. It detects the platform and distro of
selected host and uninstalls Ceph packages from it. However, some dependencies
like ``librbd1`` and ``librados2`` will not be removed because they can cause
issues with ``qemu-kvm``.
Usage::
ceph-deploy uninstall [HOST] [HOST...]
Here, [HOST] is hostname of the node from where Ceph will be uninstalled.
purge
-----
Remove Ceph packages from remote hosts and purge all data. It detects the
platform and distro of selected host, uninstalls Ceph packages and purges all
data. However, some dependencies like ``librbd1`` and ``librados2`` will not be
removed because they can cause issues with ``qemu-kvm``.
Usage::
ceph-deploy purge [HOST] [HOST...]
Here, [HOST] is hostname of the node from where Ceph will be purged.
purgedata
---------
Purge (delete, destroy, discard, shred) any Ceph data from ``/var/lib/ceph``.
Once it detects the platform and distro of desired host, it first checks if Ceph
is still installed on the selected host and if installed, it won't purge data
from it. If Ceph is already uninstalled from the host, it tries to remove the
contents of ``/var/lib/ceph``. If it fails then probably OSDs are still mounted
and needs to be unmounted to continue. It unmount the OSDs and tries to remove
the contents of ``/var/lib/ceph`` again and checks for errors. It also removes
contents of ``/etc/ceph``. Once all steps are successfully completed, all the
Ceph data from the selected host are removed.
Usage::
ceph-deploy purgedata [HOST] [HOST...]
Here, [HOST] is hostname of the node from where Ceph data will be purged.
forgetkeys
----------
Remove authentication keys from the local directory. It removes all the
authentication keys i.e, monitor keyring, client.admin keyring, bootstrap-osd
and bootstrap-mds keyring from the node.
Usage::
ceph-deploy forgetkeys
pkg
---
Manage packages on remote hosts. It is used for installing or removing packages
from remote hosts. The package names for installation or removal are to be
specified after the command. Two options :option:`--install` and
:option:`--remove` are used for this purpose.
Usage::
ceph-deploy pkg --install [PKGs] [HOST] [HOST...]
ceph-deploy pkg --remove [PKGs] [HOST] [HOST...]
Here, [PKGs] is comma-separated package names and [HOST] is hostname of the
remote node where packages are to be installed or removed from.
Options
=======
.. option:: --address
IP address of the host node to be added to the cluster.
.. option:: --adjust-repos
Install packages modifying source repos.
.. option:: --ceph-conf
Use (or reuse) a given ``ceph.conf`` file.
.. option:: --cluster
Name of the cluster.
.. option:: --dev
Install a bleeding edge built from Git branch or tag (default: master).
.. option:: --cluster-network
Specify the (internal) cluster network.
.. option:: --dmcrypt
Encrypt [data-path] and/or journal devices with ``dm-crypt``.
.. option:: --dmcrypt-key-dir
Directory where ``dm-crypt`` keys are stored.
.. option:: --install
Comma-separated package(s) to install on remote hosts.
.. option:: --fs-type
Filesystem to use to format disk ``(xfs, btrfs or ext4)``. Note that support for btrfs and ext4 is no longer tested or recommended; please use xfs.
.. option:: --fsid
Provide an alternate FSID for ``ceph.conf`` generation.
.. option:: --gpg-url
Specify a GPG key url to be used with custom repos (defaults to ceph.com).
.. option:: --keyrings
Concatenate multiple keyrings to be seeded on new monitors.
.. option:: --local-mirror
Fetch packages and push them to hosts for a local repo mirror.
.. option:: --mkfs
Inject keys to MONs on remote nodes.
.. option:: --no-adjust-repos
Install packages without modifying source repos.
.. option:: --no-ssh-copykey
Do not attempt to copy ssh keys.
.. option:: --overwrite-conf
Overwrite an existing conf file on remote host (if present).
.. option:: --public-network
Specify the public network for a cluster.
.. option:: --remove
Comma-separated package(s) to remove from remote hosts.
.. option:: --repo
Install repo files only (skips package installation).
.. option:: --repo-url
Specify a repo url that mirrors/contains Ceph packages.
.. option:: --testing
Install the latest development release.
.. option:: --username
The username to connect to the remote host.
.. option:: --version
The current installed version of :program:`ceph-deploy`.
.. option:: --zap-disk
Destroy the partition table and content of a disk.
Availability
============
:program:`ceph-deploy` is part of Ceph, a massively scalable, open-source, distributed storage system. Please refer to
the documentation at https://ceph.com/ceph-deploy/docs for more information.
See also
========
:doc:`ceph-mon <ceph-mon>`\(8),
:doc:`ceph-osd <ceph-osd>`\(8),
:doc:`ceph-volume <ceph-volume>`\(8),
:doc:`ceph-mds <ceph-mds>`\(8)