mirror of https://github.com/ceph/ceph
123 lines
4.0 KiB
ReStructuredText
123 lines
4.0 KiB
ReStructuredText
==========================================
|
|
Installing and Configuring NVMe-oF Targets
|
|
==========================================
|
|
|
|
Traditionally, block-level access to a Ceph storage cluster has been limited to
|
|
(1) QEMU and ``librbd`` (which is a key enabler for adoption within OpenStack
|
|
environments), and (2) the Linux kernel client. Starting with the Ceph Reef
|
|
release, block-level access has been expanded to offer standard NVMe/TCP
|
|
support, allowing wider platform usage and potentially opening new use cases.
|
|
|
|
Prerequisites
|
|
=============
|
|
|
|
- Red Hat Enterprise Linux/CentOS 8.0 (or newer); Linux kernel v4.16 (or newer)
|
|
|
|
- A working Ceph Reef or later storage cluster, deployed with ``cephadm``
|
|
|
|
- NVMe-oF gateways, which can either be colocated with OSD nodes or on dedicated nodes
|
|
|
|
- Separate network subnets for NVME-oF front-end traffic and Ceph back-end traffic
|
|
|
|
Explanation
|
|
===========
|
|
|
|
The Ceph NVMe-oF gateway is both an NVMe-oF target and a Ceph client. Think of
|
|
it as a "translator" between Ceph's RBD interface and the NVME-oF protocol. The
|
|
Ceph NVMe-oF gateway can run on a standalone node or be colocated with other
|
|
daemons, for example on a Ceph Object Store Disk (OSD) node. When colocating
|
|
the Ceph NVMe-oF gateway with other daemons, ensure that sufficient CPU and
|
|
memory are available. The steps below explain how to install and configure the
|
|
Ceph NVMe/TCP gateway for basic operation.
|
|
|
|
|
|
Installation
|
|
============
|
|
|
|
Complete the following steps to install the Ceph NVME-oF gateway:
|
|
|
|
#. Create a pool in which the gateways configuration can be managed:
|
|
|
|
.. prompt:: bash #
|
|
|
|
ceph osd pool create NVME-OF_POOL_NAME
|
|
|
|
#. Enable RBD on the NVMe-oF pool:
|
|
|
|
.. prompt:: bash #
|
|
|
|
rbd pool init NVME-OF_POOL_NAME
|
|
|
|
#. Deploy the NVMe-oF gateway daemons on a specific set of nodes:
|
|
|
|
.. prompt:: bash #
|
|
|
|
ceph orch apply nvmeof NVME-OF_POOL_NAME --placment="host01, host02"
|
|
|
|
Configuration
|
|
=============
|
|
|
|
Download the ``nvmeof-cli`` container before first use.
|
|
To download it use the following command:
|
|
|
|
.. prompt:: bash #
|
|
|
|
podman pull quay.io/ceph/nvmeof-cli:latest
|
|
|
|
#. Create an NVMe subsystem:
|
|
|
|
.. prompt:: bash #
|
|
|
|
podman run -it quay.io/ceph/nvmeof-cli:latest --server-address GATEWAY_IP --server-port GATEWAY_PORT 5500 subsystem add --subsystem SUSYSTEM_NQN
|
|
|
|
The subsystem NQN is a user defined string, for example ``nqn.2016-06.io.spdk:cnode1``.
|
|
|
|
#. Define the IP port on the gateway that will process the NVME/TCP commands and I/O:
|
|
|
|
a. On the install node, get the NVME-oF Gateway name:
|
|
|
|
.. prompt:: bash #
|
|
|
|
ceph orch ps | grep nvme
|
|
|
|
b. Define the IP port for the gateway:
|
|
|
|
.. prompt:: bash #
|
|
|
|
podman run -it quay.io/ceph/nvmeof-cli:latest --server-address GATEWAY_IP --server-port GATEWAY_PORT 5500 listener add --subsystem SUBSYSTEM_NQN --gateway-name GATEWAY_NAME --traddr GATEWAY_IP --trsvcid 4420
|
|
|
|
#. Get the host NQN (NVME Qualified Name) for each host:
|
|
|
|
.. prompt:: bash #
|
|
|
|
cat /etc/nvme/hostnqn
|
|
|
|
.. prompt:: bash #
|
|
|
|
esxcli nvme info get
|
|
|
|
#. Allow the initiator host to connect to the newly-created NVMe subsystem:
|
|
|
|
.. prompt:: bash #
|
|
|
|
podman run -it quay.io/ceph/nvmeof-cli:latest --server-address GATEWAY_IP --server-port GATEWAY_PORT 5500 host add --subsystem SUBSYSTEM_NQN --host "HOST_NQN1, HOST_NQN2"
|
|
|
|
#. List all subsystems configured in the gateway:
|
|
|
|
.. prompt:: bash #
|
|
|
|
podman run -it quay.io/ceph/nvmeof-cli:latest --server-address GATEWAY_IP --server-port GATEWAY_PORT 5500 subsystem list
|
|
|
|
#. Create a new NVMe namespace:
|
|
|
|
.. prompt:: bash #
|
|
|
|
podman run -it quay.io/ceph/nvmeof-cli:latest --server-address GATEWAY_IP --server-port GATEWAY_PORT 5500 namespace add --subsystem SUBSYSTEM_NQN --rbd-pool POOL_NAME --rbd-image IMAGE_NAME
|
|
|
|
#. List all namespaces in the subsystem:
|
|
|
|
.. prompt:: bash #
|
|
|
|
podman run -it quay.io/ceph/nvmeof-cli:latest --server-address GATEWAY_IP --server-port GATEWAY_PORT 5500 namespace list --subsystem SUBSYSTEM_NQN
|
|
|