================== Service Management ================== Service Status ============== A service is a group of daemons configured together. To see the status of one of the services running in the Ceph cluster, do the following: #. Use the command line to print a list of services. #. Locate the service whose status you want to check. #. Print the status of the service. The following command prints a list of services known to the orchestrator. To limit the output to services only on a specified host, use the optional ``--host`` parameter. To limit the output to services of only a particular type, use the optional ``--type`` parameter (mon, osd, mgr, mds, rgw): .. prompt:: bash # ceph orch ls [--service_type type] [--service_name name] [--export] [--format f] [--refresh] Discover the status of a particular service or daemon: .. prompt:: bash # ceph orch ls --service_type type --service_name [--refresh] To export the service specifications knows to the orchestrator, run the following command. .. prompt:: bash # ceph orch ls --export The service specifications exported with this command will be exported as yaml and that yaml can be used with the ``ceph orch apply -i`` command. For information about retrieving the specifications of single services (including examples of commands), see :ref:`orchestrator-cli-service-spec-retrieve`. Daemon Status ============= A daemon is a systemd unit that is running and part of a service. To see the status of a daemon, do the following: #. Print a list of all daemons known to the orchestrator. #. Query the status of the target daemon. First, print a list of all daemons known to the orchestrator: .. prompt:: bash # ceph orch ps [--hostname host] [--daemon_type type] [--service_name name] [--daemon_id id] [--format f] [--refresh] Then query the status of a particular service instance (mon, osd, mds, rgw). For OSDs the id is the numeric OSD ID. For MDS services the id is the file system name: .. prompt:: bash # ceph orch ps --daemon_type osd --daemon_id 0 .. _orchestrator-cli-service-spec: Service Specification ===================== A *Service Specification* is a data structure that is used to specify the deployment of services. Here is an example of a service specification in YAML: .. code-block:: yaml service_type: rgw service_id: realm.zone placement: hosts: - host1 - host2 - host3 unmanaged: false ... In this example, the properties of this service specification are: * ``service_type`` The type of the service. Needs to be either a Ceph service (``mon``, ``crash``, ``mds``, ``mgr``, ``osd`` or ``rbd-mirror``), a gateway (``nfs`` or ``rgw``), part of the monitoring stack (``alertmanager``, ``grafana``, ``node-exporter`` or ``prometheus``) or (``container``) for custom containers. * ``service_id`` The name of the service. * ``placement`` See :ref:`orchestrator-cli-placement-spec`. * ``unmanaged`` If set to ``true``, the orchestrator will not deploy nor remove any daemon associated with this service. Placement and all other properties will be ignored. This is useful, if you do not want this service to be managed temporarily. For cephadm, See :ref:`cephadm-spec-unmanaged` Each service type can have additional service-specific properties. Service specifications of type ``mon``, ``mgr``, and the monitoring types do not require a ``service_id``. A service of type ``osd`` is described in :ref:`drivegroups` Many service specifications can be applied at once using ``ceph orch apply -i`` by submitting a multi-document YAML file:: cat <. --export > rgw...yaml ceph orch ls --service-type mgr --export > mgr.yaml ceph orch ls --export > cluster.yaml The Specification can then be changed and re-applied as above. .. _orchestrator-cli-placement-spec: Placement Specification ======================= For the orchestrator to deploy a *service*, it needs to know where to deploy *daemons*, and how many to deploy. This is the role of a placement specification. Placement specifications can either be passed as command line arguments or in a YAML files. .. note:: cephadm will not deploy daemons on hosts with the ``_no_schedule`` label; see :ref:`cephadm-special-host-labels`. Explicit placements ------------------- Daemons can be explicitly placed on hosts by simply specifying them: .. prompt:: bash # orch apply prometheus --placement="host1 host2 host3" Or in YAML: .. code-block:: yaml service_type: prometheus placement: hosts: - host1 - host2 - host3 MONs and other services may require some enhanced network specifications: .. prompt:: bash # orch daemon add mon --placement="myhost:[v2:1.2.3.4:3300,v1:1.2.3.4:6789]=name" where ``[v2:1.2.3.4:3300,v1:1.2.3.4:6789]`` is the network address of the monitor and ``=name`` specifies the name of the new monitor. .. _orch-placement-by-labels: Placement by labels ------------------- Daemons can be explicitly placed on hosts that match a specific label: .. prompt:: bash # orch apply prometheus --placement="label:mylabel" Or in YAML: .. code-block:: yaml service_type: prometheus placement: label: "mylabel" * See :ref:`orchestrator-host-labels` Placement by pattern matching ----------------------------- Daemons can be placed on hosts as well: .. prompt:: bash # orch apply prometheus --placement='myhost[1-3]' Or in YAML: .. code-block:: yaml service_type: prometheus placement: host_pattern: "myhost[1-3]" To place a service on *all* hosts, use ``"*"``: .. prompt:: bash # orch apply node-exporter --placement='*' Or in YAML: .. code-block:: yaml service_type: node-exporter placement: host_pattern: "*" Setting a limit --------------- By specifying ``count``, only that number of daemons will be created: .. prompt:: bash # orch apply prometheus --placement=3 To deploy *daemons* on a subset of hosts, also specify the count: .. prompt:: bash # orch apply prometheus --placement="2 host1 host2 host3" If the count is bigger than the amount of hosts, cephadm deploys one per host: .. prompt:: bash # orch apply prometheus --placement="3 host1 host2" results in two Prometheus daemons. Or in YAML: .. code-block:: yaml service_type: prometheus placement: count: 3 Or with hosts: .. code-block:: yaml service_type: prometheus placement: count: 2 hosts: - host1 - host2 - host3 Updating Service Specifications =============================== The Ceph Orchestrator maintains a declarative state of each service in a ``ServiceSpec``. For certain operations, like updating the RGW HTTP port, we need to update the existing specification. 1. List the current ``ServiceSpec``: .. prompt:: bash # ceph orch ls --service_name= --export > myservice.yaml 2. Update the yaml file: .. prompt:: bash # vi myservice.yaml 3. Apply the new ``ServiceSpec``: .. prompt:: bash # ceph orch apply -i myservice.yaml [--dry-run] Deployment of Daemons ===================== Cephadm uses a declarative state to define the layout of the cluster. This state consists of a list of service specifications containing placement specifications (See :ref:`orchestrator-cli-service-spec` ). Cephadm constantly compares list of actually running daemons in the cluster with the desired service specifications and will either add or remove new daemons. First, cephadm will select a list of candidate hosts. It first looks for explicit host names and will select those. In case there are no explicit hosts defined, cephadm looks for a label specification. If there is no label defined in the specification, cephadm will select hosts based on a host pattern. If there is no pattern defined, cepham will finally select all known hosts as candidates. Then, cephadm will consider existing daemons of this services and will try to avoid moving any daemons. Cephadm supports the deployment of a specific amount of services. Let's consider a service specification like so: .. code-block:: yaml service_type: mds service_name: myfs placement: count: 3 label: myfs This instructs cephadm to deploy three daemons on hosts labeled with ``myfs`` across the cluster. Then, in case there are less than three daemons deployed on the candidate hosts, cephadm will then randomly choose hosts for deploying new daemons. In case there are more than three daemons deployed, cephadm will remove existing daemons. Finally, cephadm will remove daemons on hosts that are outside of the list of candidate hosts. However, there is a special cases that cephadm needs to consider. In case the are fewer hosts selected by the placement specification than demanded by ``count``, cephadm will only deploy on selected hosts. .. _cephadm-spec-unmanaged: Disable automatic deployment of daemons ======================================= Cephadm supports disabling the automated deployment and removal of daemons per service. In this case, the CLI supports two commands that are dedicated to this mode. To disable the automatic management of dameons, apply the :ref:`orchestrator-cli-service-spec` with ``unmanaged=True``. ``mgr.yaml``: .. code-block:: yaml service_type: mgr unmanaged: true placement: label: mgr .. code-block:: bash ceph orch apply -i mgr.yaml .. note:: cephadm will no longer deploy any new daemons, if the placement specification matches additional hosts. To manually deploy a daemon on a host, please execute: .. prompt:: bash # ceph orch daemon add --placement= For example : .. prompt:: bash # ceph orch daemon add mgr --placement=my_host To manually remove a daemon, please run: .. prompt:: bash # ceph orch daemon rm ... [--force] For example: .. prompt:: bash # ceph orch daemon rm mgr.my_host.xyzxyz .. note:: For managed services (``unmanaged=False``), cephadm will automatically deploy a new daemon a few seconds later. * See :ref:`cephadm-osd-declarative` for special handling of unmanaged OSDs. * See also :ref:`cephadm-pause`