mirror of https://github.com/ceph/ceph
62 lines
2.3 KiB
ReStructuredText
62 lines
2.3 KiB
ReStructuredText
|
|
||
|
CephFS for early adopters
|
||
|
=========================
|
||
|
|
||
|
This pages provides guidance for early adoption of CephFS by users
|
||
|
with an appetite for adventure. While work is ongoing to build the
|
||
|
scrubbing and disaster recovery tools needed to run CephFS in demanding
|
||
|
production environments, it is already useful for community members to
|
||
|
try CephFS and provide bug reports and feedback.
|
||
|
|
||
|
Setup instructions
|
||
|
==================
|
||
|
|
||
|
Please see the instructions at :doc:`/cephfs/index`.
|
||
|
|
||
|
Most stable configuration
|
||
|
=========================
|
||
|
|
||
|
For the best chance of a happy healthy filesystem, use a **single active MDS**
|
||
|
and **do not use snapshots**. Both of these are the default:
|
||
|
|
||
|
* Snapshots are disabled by default, unless they are enabled explicitly by
|
||
|
an administrator using the ``allow_new_snaps`` setting.
|
||
|
* Ceph will use a single active MDS unless an administrator explicitly sets
|
||
|
``max_mds`` to a value greater than 1. Note that creating additional
|
||
|
MDS daemons (e.g. with ``ceph-deploy mds create``) is okay, as these will
|
||
|
by default simply become standbys. It is also fairly safe to enable
|
||
|
standby-replay mode.
|
||
|
|
||
|
Which client?
|
||
|
=============
|
||
|
|
||
|
The fuse client is the easiest way to get up to date code, while
|
||
|
the kernel client will often give better performance.
|
||
|
|
||
|
The clients do not always provide equivalent functionality, for example
|
||
|
the fuse client supports client-enforced quotas while the kernel client
|
||
|
does not.
|
||
|
|
||
|
When encountering bugs or performance issues, it is often instructive to
|
||
|
try using the other client, in order to find out whether the bug was
|
||
|
client-specific or not (and then to let the developers know).
|
||
|
|
||
|
Reporting issues
|
||
|
================
|
||
|
|
||
|
If you have identified a specific issue, please report it with as much
|
||
|
information as possible. Especially important information:
|
||
|
|
||
|
* Ceph versions installed on client and server
|
||
|
* Whether you are using the kernel or fuse client
|
||
|
* If you are using the kernel client, what kernel version?
|
||
|
* How many clients are in play, doing what kind of workload?
|
||
|
* If a system is 'stuck', is that affecting all clients or just one?
|
||
|
* Any ceph health messages
|
||
|
* Any backtraces in the ceph logs from crashes
|
||
|
|
||
|
If you are satisfied that you have found a bug, please file it on
|
||
|
http://tracker.ceph.com. For more general queries please write
|
||
|
to the ceph-users mailing list.
|
||
|
|