mirror of
https://github.com/ceph/ceph
synced 2024-12-19 01:46:00 +00:00
6767f841e5
mon/OSDMonitor: implement cluster pg limit Reviewed-by: Josh Durgin <jdurgin@redhat.com>
32 lines
1.2 KiB
Plaintext
32 lines
1.2 KiB
Plaintext
>= 12.2.0
|
|
---------
|
|
|
|
|
|
12.2.1
|
|
------
|
|
|
|
- *RGW*
|
|
|
|
* dynamic resharding is now enabled by default, RGW will now automatically
|
|
reshard the bucket index once the index grows beyond `rgw_max_objs_per_shard`
|
|
|
|
- *CephFS*:
|
|
|
|
* Limiting MDS cache via a memory limit is now supported using the new
|
|
mds_cache_memory_limit config option (1GB by default). A cache reservation
|
|
can also be specified using mds_cache_reservation as a percentage of the
|
|
limit (5% by default). Limits by inode count are still supported using
|
|
mds_cache_size. Setting mds_cache_size to 0 (the default) disables the
|
|
inode limit.
|
|
|
|
* The maximum number of PGs per OSD before the monitor issues a
|
|
warning has been reduced from 300 to 200 PGs. 200 is still twice
|
|
the generally recommended target of 100 PGs per OSD. This limit can
|
|
be adjusted via the ``mon_max_pg_per_osd`` option on the
|
|
monitors. The older ``mon_pg_warn_max_per_osd`` option has been removed.
|
|
|
|
* Creating pools or adjusting pg_num will now fail if the change would
|
|
make the number of PGs per OSD exceed the configured
|
|
``mon_max_pg_per_osd`` limit. The option can be adjusted if it
|
|
is really necessary to create a pool with more PGs.
|