Dashboard Plugin
================
.. _mgr-dashboard-overview:
Overview
--------
The original Ceph manager dashboard that was shipped with Ceph "Luminous"
started out as a simple read-only view into various run-time information and
performance data of a Ceph cluster. It used a very simple architecture to
achieve the original goal.
However, there was a growing demand for adding more web-based management
capabilities, to make it easier to administer Ceph for users that prefer a WebUI
over using the command line.
This new dashboard module is a replacement of the previous one and an ongoing
project to add a native web based monitoring and administration application to
Ceph Manager.
The architecture and functionality of this module are derived from and inspired
by the `openATTIC Ceph management and monitoring tool
`_. The development is actively driven by the team
behind openATTIC at SUSE.
The intention is to reuse as much of the existing openATTIC functionality as
possible, while adapting it to the different environment. While openATTIC is
based on Django and the Django REST Framework, the dashboard module's backend
code uses the CherryPy framework and a custom REST API implementation.
The WebUI implementation is based on Angular/TypeScript, merging both
functionality from the original dashboard as well as adding new functionality
originally developed for the standalone version of openATTIC.
The dashboard plugin is implemented as a web application that visualizes
information and statistics about the Ceph cluster using a web server hosted by
``ceph-mgr``.
The dashboard currently provides the following features to monitor and manage
various aspects of your Ceph cluster:
* **Username/password protection**: The dashboard can only be accessed by
providing a configurable username and password.
* **SSL/TLS support**: All HTTP communication between the web browser and the
dashboard is secured via SSL. A self-signed certificate can be created with
a built-in command, but it's also possible to import custom certificates
signed and issued by a CA.
* **Overall cluster health**: Displays the overall cluster status, storage
utilization (e.g. number of objects, raw capacity, usage per pool), a list of
pools and their status and usage statistics.
* **Cluster logs**: Display the latest updates to the cluster's event and audit
log files.
* **Hosts**: Provides a list of all hosts associated to the cluster, which
services are running and which version of Ceph is installed.
* **Performance counters**: Displays detailed service-specific statistics for
each running service.
* **Monitors**: Lists all MONs, their quorum status, open sessions.
* **Configuration Reference**: Lists all available configuration options,
their description and default values.
* **Pools**: List all Ceph pools and their details (e.g. applications, placement
groups, replication size, EC profile, CRUSH ruleset, etc.)
* **OSDs**: Lists all OSDs, their status and usage statistics as well as
detailed information like attributes (OSD map), metadata, performance counters
and usage histograms for read/write operations.
* **iSCSI**: Lists all hosts that run the TCMU runner service, displaying all
images and their performance characteristics (read/write ops, traffic).
* **RBD**: Lists all RBD images and their properties (size, objects, features).
Create, copy, modify and delete RBD images. Create, delete and rollback
snapshots of selected images, protect/unprotect these snapshots against
modification. Copy or clone snapshots, flatten cloned images.
* **RBD mirroring**: Lists all active sync daemons and their status, pools and
RBD images including their synchronization state.
* **CephFS**: Lists all active filesystem clients and associated pools,
including their usage statistics.
* **Object Gateway**: Lists all active object gateways and their performance
counters. Display and manage (add/edit/delete) object gateway users and their
details (e.g. quotas) as well as the users' buckets and their details (e.g.
owner, quotas).
Enabling
--------
Within a running Ceph cluster, the *dashboard* manager module is enabled with::
$ ceph mgr module enable dashboard
This can be automated (e.g. during deployment) by adding the following to
``ceph.conf``::
[mon]
mgr initial modules = dashboard
Note that ``mgr initial modules`` takes a space-separated list of modules, so
if you wanted to include other modules in addition to dashboard, just make it
a list like so::
mgr initial modules = balancer dashboard
Configuration
-------------
SSL/TLS Support
^^^^^^^^^^^^^^^
All HTTP connections to the dashboard are secured with SSL/TLS.
To get the dashboard up and running quickly, you can generate and install a
self-signed certificate using the following built-in command::
$ ceph dashboard create-self-signed-cert
Note that most web browsers will complain about such self-signed certificates
and require explicit confirmation before establishing a secure connection to the
dashboard.
To properly secure a deployment and to remove the certificate warning, a
certificate that is issued by a certificate authority (CA) should be used.
For example, a key pair can be generated with a command similar to::
$ openssl req -new -nodes -x509 \
-subj "/O=IT/CN=ceph-mgr-dashboard" -days 3650 \
-keyout dashboard.key -out dashboard.crt -extensions v3_ca
The ``dashboard.crt`` file should then be signed by a CA. Once that is done, you
can enable it for all Ceph manager instances by running the following commands::
$ ceph config-key set mgr mgr/dashboard/crt -i dashboard.crt
$ ceph config-key set mgr mgr/dashboard/key -i dashboard.key
If different certificates are desired for each manager instance for some reason,
the name of the instance can be included as follows (where ``$name`` is the name
of the ``ceph-mgr`` instance, usually the hostname)::
$ ceph config-key set mgr/dashboard/$name/crt -i dashboard.crt
$ ceph config-key set mgr/dashboard/$name/key -i dashboard.key
.. note::
You need to restart the Ceph manager processes manually after changing the SSL
certificate and key. This can be accomplished by either running ``ceph mgr
fail mgr`` or by disabling and re-enabling the dashboard module (which also
triggers the manager to respawn itself)::
$ ceph mgr module disable dashboard
$ ceph mgr module enable dashboard
Host name and port
^^^^^^^^^^^^^^^^^^
Like most web applications, dashboard binds to a TCP/IP address and TCP port.
By default, the ``ceph-mgr`` daemon hosting the dashboard (i.e., the currently
active manager) will bind to TCP port 8080. If no specific address has been
configured, the web app will bind to ``::``, which corresponds to all available
IPv4 and IPv6 addresses.
These defaults can be changed via the configuration key facility on a
cluster-wide level (so they apply to all manager instances) as follows::
$ ceph config set mgr mgr/dashboard/server_addr $IP
$ ceph config set mgr mgr/dashboard/server_port $PORT
Since each ``ceph-mgr`` hosts its own instance of dashboard, it may also be
necessary to configure them separately. The IP address and port for a specific
manager instance can be changed with the following commands::
$ ceph config set mgr mgr/dashboard/$name/server_addr $IP
$ ceph config set mgr mgr/dashboard/$name/server_port $PORT
Replace ``$name`` with the ID of the ceph-mgr instance hosting the dashboard web
app.
.. note::
The command ``ceph mgr services`` will show you all endpoints that are
currently configured. Look for the "dashboard" key to obtain the URL for
accessing the dashboard.
Username and password
^^^^^^^^^^^^^^^^^^^^^
In order to be able to log in, you need to define a username and password, which
will be stored in the MON's configuration database::
$ ceph dashboard set-login-credentials
The password will be stored in the configuration database in encrypted form
using ``bcrypt``. This is a global setting that applies to all dashboard instances.
Enabling the Object Gateway management frontend
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
To use the Object Gateway management functionality of the dashboard, you will
need to provide the login credentials of a user with the ``system`` flag
enabled.
If you do not have a user which shall be used for providing those credentials,
you will also need to create one::
$ radosgw-admin user create --uid= --display-name= \
--system
Take note of the keys ``access_key`` and ``secret_key`` in the output of this
command.
The credentials of an existing user can also be obtained by using
`radosgw-admin`::
$ radosgw-admin user info --uid=
Finally, provide the credentials to the dashboard module::
$ ceph dashboard set-rgw-api-access-key
$ ceph dashboard set-rgw-api-secret-key
This is all you have to do to get the Object Gateway management functionality
working. The host and port of the Object Gateway are determined automatically.
If multiple zones are used, it will automatically determine the host within the
master zone group and master zone. This should be sufficient for most setups,
but in some circumstances you might want to set the host and port manually::
$ ceph dashboard set-rgw-api-host
$ ceph dashboard set-rgw-api-port
In addition to the settings mentioned so far, the following settings do also
exist and you may find yourself in the situation that you have to use them::
$ ceph dashboard set-rgw-api-scheme # http or https
$ ceph dashboard set-rgw-api-admin-resource
$ ceph dashboard set-rgw-api-user-id
Enabling the Embedding of Grafana Dashboards
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
Grafana and Prometheus are likely going to be bundled and installed by some
orchestration tools along Ceph in the near future, but currently, you will have
to install and configure both manually. After you have installed Prometheus and
Grafana on your preferred hosts, proceed with the following steps::
1. Enable the Ceph Exporter which comes as Ceph Manager module by running::
$ ceph mgr module enable prometheus
More details can be found on the `documentation
`_ of the prometheus
module.
2. Add the corresponding scrape configuration to Prometheus. This may look
like::
global:
scrape_interval: 5s
scrape_configs:
- job_name: 'prometheus'
static_configs:
- targets: ['localhost:9090']
- job_name: 'ceph'
static_configs:
- targets: ['localhost:9283']
- job_name: 'node-exporter'
static_configs:
- targets: ['localhost:9100']
3. Add Prometheus as data source to Grafana
4. Install the `vonage-status-panel` plugin using::
grafana-cli plugins install vonage-status-panel
4. Add the Dashboards to Grafana by importing them
5. Configure Grafana in `/etc/grafana/grafana.ini` to adapt the URLs to the
Ceph Dashboard properly::
root_url = http://localhost:3000/api/grafana/proxy
After you have configured Grafana and Prometheus, you will need to tell the
Ceph Manager Dashboard where it can access Grafana and what the credentials are
to do so. This can be done by using the following commands::
$ ceph dashboard set-grafana-api-url # default: 'http://localhost:3000'
$ ceph dashboard set-grafana-api-username # default: 'admin'
$ ceph dashboard set-grafana-api-password # default: 'admin'
Accessing the dashboard
^^^^^^^^^^^^^^^^^^^^^^^
You can now access the dashboard using your (JavaScript-enabled) web browser, by
pointing it to any of the host names or IP addresses and the selected TCP port
where a manager instance is running: e.g., ``httpS://<$IP>:<$PORT>/``.
You should then be greeted by the dashboard login page, requesting your
previously defined username and password. Select the **Keep me logged in**
checkbox if you want to skip the username/password request when accessing the
dashboard in the future.
Reverse proxies
---------------
If you are accessing the dashboard via a reverse proxy configuration,
you may wish to service it under a URL prefix. To get the dashboard
to use hyperlinks that include your prefix, you can set the
``url_prefix`` setting:
::
ceph config set mgr mgr/dashboard/url_prefix $PREFIX
so you can access the dashboard at ``http://$IP:$PORT/$PREFIX/``.