Remove the misleading reference to this from the client
eviction page, it was never the right option to mention
there (my mistake).
Demote the option from LEVEL_ADVANCED to LEVEL_DEV as it
is hard to imagine a good reason for the user to change it.
Set a hard minimum of one hour, to make it harder to
corrupt` a system by setting it close to zero.
Remove the legacy definition of the field while we're at it.
Fixes: http://tracker.ceph.com/issues/21821
Signed-off-by: John Spray <john.spray@redhat.com>
The "mds blacklist interval" setting has no effect on the time that
the "ceph osd blacklist" command will use by default. Clarify this in
the docs.
Signed-off-by: Ken Dreyer <kdreyer@redhat.com>
* refs/remotes/upstream/pull/16036/head:
mds: improve cap min/max ratio descriptions
mds: fix whitespace
mds: cap client recall to min caps per client
mds: fix conf types
mds: fix whitespace
doc/cephfs: add client min cache and max cache ratio describe
mds: adding tunable features for caps_per_client
Reviewed-by: Patrick Donnelly <pdonnell@redhat.com>
Reviewed-by: Zheng Yan <zyan@redhat.com>
* refs/remotes/upstream/pull/17657/head:
mds: optimize MDCache::rejoin_scour_survivor_replicas()
mds: fix MDSCacheObject::clear_replica_map
mds: support limiting cache by memory
common: refactor of lru
mds: resolve unsigned coercion compiler warning
common: use safer uint64_t for list size
common: add bytes2str pretty print function
mds: check if waiting is allocated before use
mds: go back to compact_map for replicas
mds: use mempool for cache objects
mds: cleanup replica_map access
common: add alloc_ptr smart pointer
common: add warning on base class use of mempool
common: use atomic uin64_t for counter
Reviewed-by: Zheng Yan <zyan@redhat.com>
* refs/remotes/upstream/pull/17608/head:
doc/cephfs/posix: put posix notes in perspective
Reviewed-by: John Spray <john.spray@redhat.com>
Reviewed-by: Patrick Donnelly <pdonnell@redhat.com>
This introduces two config parameters:
mds_cache_memory_limit: Sets the soft maximum of the cache to the given
byte count. (Like mds_cache_size, this doesn't actually limit the maximum
size of the cache. It just dictates the steady-state size.)
mds_cache_reservation: This replaces mds_health_cache_threshold everywhere
except the Beacon heartbeat sent to the mons. The idea here is to specify a
reservation of memory (5% by default) for operations and the MDS tries to
always maintain that reservation. So, the MDS will recall caps from clients
when it begins dipping into its reservation of memory.
mds_cache_size still limits the cache by Inode count but is now by-default 0
(i.e. unlimited). The new preferred way of specifying cache limits is by memory
size. The default is 1GB.
Fixes: http://tracker.ceph.com/issues/20594
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1464976
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
Current cephfs can support seekdir efficiently. The diverge was
fixed by https://github.com/ceph/ceph/pull/14317
Signed-off-by: "Yan, Zheng" <zyan@redhat.com>
Add a short description and command example to set the allow_multimds
flag and add a <fs_name> place holder to all 'ceph fs set' commands.
Signed-off-by: Jan Fajerski <jfajerski@suse.com>
Add a description of max_file_size to the CephFS admin docs.
Thanks to John Spray <jspray@redhat.com> on ceph-users for this
information.
Signed-off-by: Ken Dreyer <kdreyer@redhat.com>
The "session ls" and "session evict" are now
"client ls" and "client evict" (the old ones
are still there for backwards compatibility).
The automatic client eviction now emits
cluster logs that call the client by its
friendly name (usually the hostname).
Signed-off-by: John Spray <john.spray@redhat.com>
This makes the code a bit awkward because we now have
4x paths through this function (blocking, noblocking, blacklisting,
nonblacklisting).
Signed-off-by: John Spray <john.spray@redhat.com>
This was occasionally useful for establishing a journal-less
performance baseline, but it has two big problems:
* We don't test it, so it's probably broken a lot of the time
* It sounds a lot to a naive user like an option for controlling
logging.
IMO, anyone who wants this behaviour is in sufficiently advanced
territory that then can hack it in and recompile, we don't need
to leave dangerous things like this in our releases.
Fixes: http://tracker.ceph.com/issues/18816
Signed-off-by: John Spray <john.spray@redhat.com>
This allows the client/admin to pin a directory tree to a particular rank,
preventing its export by the dynamic balancer.
Fixes: http://tracker.ceph.com/issues/17834
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
This had been broken for some time, as since the new
JournalStream stuff, zero padding was no longer a valid
encoding.
Fixes: http://tracker.ceph.com/issues/19691
Signed-off-by: John Spray <john.spray@redhat.com>
Client.cc marks session as stale instead of reconecting after received
reset from MDS. On MDS side session is closed so MDS is ignoring cap
renew. This adds option to reconnect stale client sessions instead of
just marking sessions stale.
Fixes: http://tracker.ceph.com/issues/18757
Signed-off-by: Henrik Korkuc <henrik@kirneh.eu>
Add a procedure that permits reconstructing metadata in a potentially
damaged cephfs metadata pool and writing the results into a
freshly-initialized pool that refers to the same data pool. Add option
flags to override checks that would ordinarily prevent this and add
options to the recovery tools to write output to a separate pool instead of
the one selected for recovery. See docs/cephfs/disaster-recovery.rst for
details.
Fixes: http://tracker.ceph.com/issues/15068
Fixes: http://tracker.ceph.com/issues/15069
Signed-off-by: Douglas Fuller <dfuller@redhat.com>
Rewrote mount.fuse.ceph to move ceph-fuse options to `fs_mntops`, where it should be.
Bash version with options in `fs_spec` is counterintuitive and cause issues in some
situations (systemd, for example)
Signed-off-by: Edgaras Lukosevicius <edgaras.lukosevicius@gmail.com>
As described in http://tracker.ceph.com/issues/17937, a client with
restricted pool access can still delete files unless a corresponding
MDS path restriction is also in place.
Signed-off-by: David Disseldorp <ddiss@suse.de>
The "additional details" section was getting longer
and longer, and I don't want developer-only stuff
in there anyway.
Signed-off-by: John Spray <john.spray@redhat.com>
Some files had multiple top-level headers resulting in
subsections being listed in the TOC alongside the
main page header.
Signed-off-by: John Spray <john.spray@redhat.com>
...and add a warning at the very top of the document
that it is not for ordinary user use.
Considered just not linking this in at all, but it
seems a shame to have it here but not be able to
e.g. google it and link to it.
Signed-off-by: John Spray <john.spray@redhat.com>
Sometimes users know that particular data pool PGs
have been damaged, and they would like to scan
their files to work out which ones might have
been affected.
Fixes: http://tracker.ceph.com/issues/17249
Signed-off-by: John Spray <john.spray@redhat.com>
It is now required that all changes to max_mds use the run-time `ceph fs
set max_mds` command. The rationale for this change is that it is
confusing to have a configuration for max_mds which is only observed at
file system creation.
Fixes: http://tracker.ceph.com/issues/17105
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
When testing < v10.2.1 with standby replay, I was often confused why the
MDS daemons would go into replay even when this was false.
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
I tried using these settings in tests without success. The correct config names
are prefixed with "mds".
Fixes: http://tracker.ceph.com/issues/16664
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
...and restructure the "early adopters" page
into a "best practices" guide. Early adopters
are now just adopters :-)
Signed-off-by: John Spray <john.spray@redhat.com>
When user is mounted a quota-restricted inode
as the root, report that inode's quota status
as the filesystem statistics in statfs.
This allows us to have a fairly convincing illusion
that someone has a filesystem to themselves, when
they're really mounting a restricted part of
the larger global filesystem.
Fixes: http://tracker.ceph.com/issues/15599
Signed-off-by: John Spray <john.spray@redhat.com>
Principally to tell people how to configure
standby, but more generally as a place to define
our terms for GIDs etc.
Signed-off-by: John Spray <john.spray@redhat.com>
As 'fs remove_data_pool' is newly introduced command, let's
change it to 'fs rm_data_pool' immediately rather than introduce
another alias for it.
Signed-off-by: Dongsheng Yang <dongsheng.yang@easystack.cn>
Creating this skeleton to include the new
"ceph fs ..." commands. Needs filling out
with prose about purpose and result of
commands.
Signed-off-by: John Spray <john.spray@redhat.com>
These are deliberately fairly sparse, because:
* These tools are for experts
* These tools may well be wrapped in a higher
level recovery tool that orchestrates parallel
workers at some stage.
Signed-off-by: John Spray <john.spray@redhat.com>