====================== Ceph Storage Cluster ====================== The :term:`Ceph Storage Cluster` is the foundation for all Ceph deployments. Based upon :abbr:`RADOS (Reliable Autonomic Distributed Object Store)`, Ceph Storage Clusters consist of two types of daemons: a :term:`Ceph OSD Daemon` (OSD) stores data as objects on a storage node; and a :term:`Ceph Monitor` (MON) maintains a master copy of the cluster map. A Ceph Storage Cluster may contain thousands of storage nodes. A minimal system will have at least one Ceph Monitor and two Ceph OSD Daemons for data replication. The Ceph File System, Ceph Object Storage and Ceph Block Devices read data from and write data to the Ceph Storage Cluster. .. raw:: html

Config and Deploy

Ceph Storage Clusters have a few required settings, but most configuration settings have default values. A typical deployment uses a deployment tool to define a cluster and bootstrap a monitor. See `Deployment`_ for details on ``cephadm.`` .. toctree:: :maxdepth: 2 Configuration Deployment <../cephadm/index> .. raw:: html

Operations

Once you have deployed a Ceph Storage Cluster, you may begin operating your cluster. .. toctree:: :maxdepth: 2 Operations .. toctree:: :maxdepth: 1 Man Pages .. toctree:: :hidden: troubleshooting/index .. raw:: html

APIs

Most Ceph deployments use `Ceph Block Devices`_, `Ceph Object Storage`_ and/or the `Ceph File System`_. You may also develop applications that talk directly to the Ceph Storage Cluster. .. toctree:: :maxdepth: 2 APIs .. raw:: html
.. _Ceph Block Devices: ../rbd/ .. _Ceph File System: ../cephfs/ .. _Ceph Object Storage: ../radosgw/ .. _Deployment: ../cephadm/