Ceph is a distributed object, block, and file storage platform
Go to file
Adam King b3b6577e24
Merge pull request #61783 from adk3798/cephadm-nvmeof-one-daemon-per-node
mgr/cephadm: block deploying nvmeof daemons of different groups on same host

Reviewed-by: John Mulligan <jmulligan@redhat.com>
2025-03-10 15:22:08 -04:00
.github Merge pull request #62019 from ceph/template-updates 2025-02-27 12:04:43 -05:00
admin doc: Upgrade and unpin some python versions 2025-02-12 18:46:06 -05:00
bin scripts: fix bash path in shebangs (part 2) 2017-09-25 17:20:40 -06:00
ceph-erasure-code-corpus@2d7d78b9cc tests: erasure-code: add k=2, m=2 for isa & jerasure 2016-12-08 12:44:21 +01:00
ceph-menv ceph-menv:fix typo in README 2023-11-17 12:39:14 +08:00
ceph-object-corpus@6b15dbab56 ceph-object-corpus: update the submodule 2024-12-10 06:23:29 +00:00
cmake/modules cmake/modules/BuildISAL.cmake: set no-integrated-as on clang only 2025-03-02 08:48:31 +00:00
container container/build.sh: remove local container images 2025-02-26 16:17:27 -08:00
debian test/rgw/lua: install luarocks 2025-01-16 14:57:09 +00:00
doc doc/releases: Add ordering comment to releases.yml 2025-03-08 10:58:37 -05:00
etc packaging: remove SuSEfirewall2 support 2019-07-12 09:55:35 +02:00
examples examples/rgw: register boto extension for BucketIndex on CreateBucket 2025-02-25 11:15:57 -05:00
fusetrace include/unordered_{map,set}: remove, use std directly instead 2025-02-20 06:12:10 +01:00
keys new release key 2015-09-17 12:28:25 -04:00
man doc,man: typos found by codespell 2021-12-15 12:04:36 +01:00
mirroring doc: Remove inactive Ceph mirrors. 2022-08-19 09:47:26 +08:00
monitoring monitoring: fix NVMeoFSubsystemNamespaceLimit 2025-02-26 21:31:54 +05:30
qa Merge pull request #62189 from idryomov/wip-66419 2025-03-10 17:02:08 +01:00
selinux selinux: prepare for anon inode controls enablement 2022-08-29 14:42:44 +02:00
share ceph-post-file: migrate to RSA SSH keys 2016-08-22 10:27:36 -04:00
src Merge pull request #61783 from adk3798/cephadm-nvmeof-one-daemon-per-node 2025-03-10 15:22:08 -04:00
sudoers.d rpm, debian: move smartmontools and nvme-cli to ceph-base 2021-08-25 13:20:23 +00:00
systemd Merge pull request #55560 from tchaikov/systemd-sans-clsuter-name 2024-05-08 21:40:36 +08:00
udev rbd: include RADOS namespace in krbd symlinks 2020-09-08 19:36:02 +02:00
.gitattributes doc: use union merge strategy for resolving conflicts in PendingReleaseNotes 2020-08-28 18:59:30 -07:00
.githubmap Merge PR #62108 into main 2025-03-06 08:38:52 -05:00
.gitignore Revert "pybind/mgr: Protobuffer file generation for grpc deps." 2025-03-04 20:05:10 +05:30
.gitleaks.toml .gitleaks.toml: Allowlist cephadm test 2022-08-08 13:57:36 -04:00
.gitmodule_mirrors Replace gmock/gtest with googletest 2016-07-12 17:18:09 -07:00
.gitmodules .gitmodules: remove shallow=true config from nvmeof/gateway 2024-08-20 18:25:12 -07:00
.mailmap Add Abhishek Kane into peoplemap, githubmap, organizationmap, mailmap 2025-01-16 11:36:06 +05:30
.organizationmap Add Abhishek Kane into peoplemap, githubmap, organizationmap, mailmap 2025-01-16 11:36:06 +05:30
.peoplemap Add Abhishek Kane into peoplemap, githubmap, organizationmap, mailmap 2025-01-16 11:36:06 +05:30
.readthedocs.yml doc: Upgrade and unpin some python versions 2025-02-12 18:46:06 -05:00
AUTHORS AUTHORS: Update Component Technical Leads and maintainers to canonical location of ceph.com. 2017-10-18 18:48:37 +08:00
ceph.spec.in Merge pull request #61098 from anoopcs9/add-libcephfs-daemon-containerfile 2025-01-21 10:37:30 -08:00
CMakeLists.txt cmake/cephfs: fix options to enable client and dependencies 2025-03-05 13:23:18 -05:00
CodingStyle CodingStyle: allow C++ forward declarations 2024-10-15 18:10:46 +02:00
ContainerBuild.md doc: document the new container build tool and link to it in README 2025-02-28 15:16:35 -05:00
CONTRIBUTING.rst doc/rados/configuration/common.rst: expand the running multiple clusters section 2020-11-14 01:36:07 -08:00
COPYING Merge pull request #59838 from cbodley/wip-68083 2024-10-30 11:14:22 -07:00
COPYING-GPL2 Add GPLv2 text file 2015-02-27 10:43:26 -07:00
COPYING-LGPL2.1 COPYING: note licenses for all files, not just the default 2012-01-12 10:03:27 -08:00
COPYING-LGPL3 relicense LGPL-2.1 code as LGPL-2.1 or LGPL-3.0 2019-04-22 11:22:55 -05:00
do_cmake.sh do_cmake.sh: specify correct python version for centos 10 stream 2025-02-20 16:26:37 -05:00
do_freebsd.sh do_freebsd.sh: set WITH_JAEGER=OFF 2022-05-17 04:42:01 -04:00
doc_deps.deb.txt admin/build-doc: s/virtualenv/python3 -m venv/ 2021-07-31 22:34:05 +08:00
Dockerfile.build Dockerfile.build: Install sccache 2025-03-05 13:40:03 -07:00
install-deps.sh CMakeLists: add boost::locale dependency for client 2025-02-27 13:41:56 -05:00
make-debs.sh build: small script tweak to allow different build dirs 2025-02-05 15:13:06 -05:00
make-dist make-dist: Quiet submodule operations in Jenkins 2024-11-08 12:43:19 -07:00
make-srpm.sh make-srpm.sh: Skip make-dist if tarball is present 2025-03-05 13:39:27 -07:00
mingw_conf.sh common: Windows Unicode CLI support 2023-11-22 09:14:49 +00:00
package-lock.json mgr/dashboard: fixed unit tests 2025-02-06 16:09:38 +00:00
PendingReleaseNotes Merge PR #60746 into main 2025-03-03 08:31:36 -05:00
pom.xml update license for libcephfs 2014-08-26 15:52:37 -04:00
README.aix ceph.spec.in: drop gdbm from build deps 2021-08-18 01:04:42 +08:00
README.FreeBSD doc: fix typos 2018-09-21 12:41:42 +08:00
README.md doc: document the new container build tool and link to it in README 2025-02-28 15:16:35 -05:00
README.solaris aix gcc librados port 2015-11-23 00:03:16 -05:00
README.windows.rst win32_build.sh: use ENABLE_SHARED=ON by default 2023-09-06 11:17:16 +00:00
run-make-check.sh run-make-check: Enable ctest resource allocation 2024-02-21 10:36:21 +08:00
SECURITY.md SECURITY.md: update security policy 2021-05-14 19:11:02 +05:30
SubmittingPatches-backports.rst doc: SubmittingPatches-backports - remove backports team 2024-10-10 22:11:22 +10:00
SubmittingPatches-kernel.rst doc: split up SubmittingPatches.rst 2019-10-17 19:32:03 +02:00
SubmittingPatches.rst SubmittingPatches.rst: PRs should target "main" 2022-06-13 09:38:40 +08:00
win32_build.sh win32: add libicu Windows build 2025-02-27 13:41:57 -05:00
win32_deps_build.sh win32: add libicu Windows build 2025-02-27 13:41:57 -05:00

Ceph - a scalable distributed storage system

See https://ceph.com/ for current information about Ceph.

Status

OpenSSF Best Practices Issue Backporting

Contributing Code

Most of Ceph is dual-licensed under the LGPL version 2.1 or 3.0. Some miscellaneous code is either public domain or licensed under a BSD-style license.

The Ceph documentation is licensed under Creative Commons Attribution Share Alike 3.0 (CC-BY-SA-3.0).

Some headers included in the ceph/ceph repository are licensed under the GPL. See the file COPYING for a full inventory of licenses by file.

All code contributions must include a valid "Signed-off-by" line. See the file SubmittingPatches.rst for details on this and instructions on how to generate and submit patches.

Assignment of copyright is not required to contribute code. Code is contributed under the terms of the applicable license.

Checking out the source

Clone the ceph/ceph repository from github by running the following command on a system that has git installed:

git clone git@github.com:ceph/ceph

Alternatively, if you are not a github user, you should run the following command on a system that has git installed:

git clone https://github.com/ceph/ceph.git

When the ceph/ceph repository has been cloned to your system, run the following commands to move into the cloned ceph/ceph repository and to check out the git submodules associated with it:

cd ceph
git submodule update --init --recursive --progress

Build Prerequisites

section last updated 06 Sep 2024

We provide the Debian and Ubuntu apt commands in this procedure. If you use a system with a different package manager, then you will have to use different commands.

#. Install curl:

apt install curl

#. Install package dependencies by running the install-deps.sh script:

./install-deps.sh

#. Install the python3-routes package:

apt install python3-routes

Building Ceph

These instructions are meant for developers who are compiling the code for development and testing. To build binaries that are suitable for installation we recommend that you build .deb or .rpm packages, or refer to ceph.spec.in or debian/rules to see which configuration options are specified for production builds.

To build Ceph, follow this procedure:

  1. Make sure that you are in the top-level ceph directory that contains do_cmake.sh and CONTRIBUTING.rst.

  2. Run the do_cmake.sh script:

    ./do_cmake.sh
    

    do_cmake.sh by default creates a "debug build" of Ceph, which can be up to five times slower than a non-debug build. Pass -DCMAKE_BUILD_TYPE=RelWithDebInfo to do_cmake.sh to create a non-debug build.

  3. Move into the build directory:

    cd build
    
  4. Use the ninja buildsystem to build the development environment:

    ninja -j3
    

    Important

    Ninja is the build system used by the Ceph project to build test builds. The number of jobs used by ninja is derived from the number of CPU cores of the building host if unspecified. Use the -j option to limit the job number if build jobs are running out of memory. If you attempt to run ninja and receive a message that reads g++: fatal error: Killed signal terminated program cc1plus, then you have run out of memory.

    Using the -j option with an argument appropriate to the hardware on which the ninja command is run is expected to result in a successful build. For example, to limit the job number to 3, run the command ninja -j3. On average, each ninja job run in parallel needs approximately 2.5 GiB of RAM.

    This documentation assumes that your build directory is a subdirectory of the ceph.git checkout. If the build directory is located elsewhere, point CEPH_GIT_DIR to the correct path of the checkout. Additional CMake args can be specified by setting ARGS before invoking do_cmake.sh. See cmake options for more details. For example:

    ARGS="-DCMAKE_C_COMPILER=gcc-7" ./do_cmake.sh
    

    To build only certain targets, run a command of the following form:

    ninja [target name]
    
  5. Install the vstart cluster:

    ninja install
    

CMake Options

The -D flag can be used with cmake to speed up the process of building Ceph and to customize the build.

Building without RADOS Gateway

The RADOS Gateway is built by default. To build Ceph without the RADOS Gateway, run a command of the following form:

cmake -DWITH_RADOSGW=OFF [path to top-level ceph directory]

Building with debugging and arbitrary dependency locations

Run a command of the following form to build Ceph with debugging and alternate locations for some external dependencies:

cmake -DCMAKE_INSTALL_PREFIX=/opt/ceph -DCMAKE_C_FLAGS="-Og -g3 -gdwarf-4" \
..

Ceph has several bundled dependencies such as Boost, RocksDB and Arrow. By default, cmake builds these bundled dependencies from source instead of using libraries that are already installed on the system. You can opt to use these system libraries, as long as they meet Ceph's version requirements. To use system libraries, use cmake options like WITH_SYSTEM_BOOST, as in the following example:

cmake -DWITH_SYSTEM_BOOST=ON [...]

To view an exhaustive list of -D options, invoke cmake -LH:

cmake -LH

Preserving diagnostic colors

If you pipe ninja to less and would like to preserve the diagnostic colors in the output in order to make errors and warnings more legible, run the following command:

cmake -DDIAGNOSTICS_COLOR=always ...

The above command works only with supported compilers.

The diagnostic colors will be visible when the following command is run:

ninja | less -R

Other available values for DIAGNOSTICS_COLOR are auto (default) and never.

Tips and Tricks

  • Use "debug builds" only when needed. Debugging builds are helpful for development, but they can slow down performance. Use -DCMAKE_BUILD_TYPE=Release when debugging isn't necessary.
  • Enable Selective Daemons when testing specific components. Don't start unnecessary daemons.
  • Preserve Existing Data skip cluster reinitialization between tests by using the -n flag.
  • To manage a vstart cluster, stop daemons using ./stop.sh and start them with ./vstart.sh --daemon osd.${ID} [--nodaemonize].
  • Restart the sockets by stopping and restarting the daemons associated with them. This ensures that there are no stale sockets in the cluster.
  • To track RocksDB performance, set export ROCKSDB_PERF=true and start the cluster by using the command ./vstart.sh -n -d -x --bluestore.
  • Build with vstart-base using debug flags in cmake, compile, and deploy via ./vstart.sh -d -n --bluestore.
  • To containerize, generate configurations with vstart.sh, and deploy with Docker, mapping directories and configuring the network.
  • Manage containers using docker run, stop, and rm. For detailed setups, consult the Ceph-Container repository.

Troubleshooting

  • Cluster Fails to Start: Look for errors in the logs under the out/ directory.
  • OSD Crashes: Check the OSD logs for errors.
  • Cluster in a Health Error State: Run the ceph status command to identify the issue.
  • RocksDB Errors: Look for RocksDB-related errors in the OSD logs.

Building a source tarball

To build a complete source tarball with everything needed to build from source and/or build a (deb or rpm) package, run

./make-dist

This will create a tarball like ceph-$version.tar.bz2 from git. (Ensure that any changes you want to include in your working directory are committed to git.)

Running a test cluster

From the ceph/ directory, run the following commands to launch a test Ceph cluster:

cd build
ninja vstart        # builds just enough to run vstart
../src/vstart.sh --debug --new -x --localhost --bluestore
./bin/ceph -s

Most Ceph commands are available in the bin/ directory. For example:

./bin/rbd create foo --size 1000
./bin/rados -p foo bench 30 write

To shut down the test cluster, run the following command from the build/ directory:

../src/stop.sh

Use the sysvinit script to start or stop individual daemons:

./bin/init-ceph restart osd.0
./bin/init-ceph stop

Running unit tests

To build and run all tests (in parallel using all processors), use ctest:

cd build
ninja
ctest -j$(nproc)

(Note: Many targets built from src/test are not run using ctest. Targets starting with "unittest" are run in ninja check and thus can be run with ctest. Targets starting with "ceph_test" can not, and should be run by hand.)

When failures occur, look in build/Testing/Temporary for logs.

To build and run all tests and their dependencies without other unnecessary targets in Ceph:

cd build
ninja check -j$(nproc)

To run an individual test manually, run ctest with -R (regex matching):

ctest -R [regex matching test name(s)]

(Note: ctest does not build the test it's running or the dependencies needed to run it)

To run an individual test manually and see all the tests output, run ctest with the -V (verbose) flag:

ctest -V -R [regex matching test name(s)]

To run tests manually and run the jobs in parallel, run ctest with the -j flag:

ctest -j [number of jobs]

There are many other flags you can give ctest for better control over manual test execution. To view these options run:

man ctest

Building Ceph using Containers

Ceph now provides tools to build the code, run unit tests, or build packages from within an OCI-style container using Podman or Docker! This allows one to build code for distributions other than the one you have on your system, avoids the need to install build dependencies for Ceph on your local system and provides an opportunity to test builds on platforms that are not yet supported by the official build infrastructure. For more details see the container build document.

Building the Documentation

Prerequisites

The list of package dependencies for building the documentation can be found in doc_deps.deb.txt:

sudo apt-get install `cat doc_deps.deb.txt`

Building the Documentation

To build the documentation, ensure that you are in the top-level /ceph directory, and execute the build script. For example:

admin/build-doc

Reporting Issues

To report an issue and view existing issues, please visit https://tracker.ceph.com/projects/ceph.