Include test case
Configurable by setting mon_osd_warn_num_repaired (default 10)
Ignore new health warning with random eio injection test
Fixes: https://tracker.ceph.com/issues/41564
Signed-off-by: David Zafman <dzafman@redhat.com>
CVEs fixed before the first stable Pacific release need not be mentioned
explicitly in the release notes.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
The RGW "default number of bucket index shards for new buckets was raised from
1 to 11" release note was added to PendingReleaseNotes by
1ec77ec7ba but it did not find its way into the
official v15.2.0 release notes until now.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
The RGW BlockPublicAccess release note was added to PendingReleaseNotes
by df384ea95f but it did not find its way
into the official v15.2.0 release notes until now.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
The RGW bucket granularity release note was added to PendingReleaseNotes
by 50f13fd54b but it never found its way
into the official v15.2.0 release notes.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
The "enforce mds_join_fs cluster affinity" release note was
added to PendingReleaseNotes by 3fac3b1236
but it never made its way into the official v15.2.0 release notes.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
The release note about osd_calc_pg_upmaps_max_stddev removal
was added to PendingReleaseNotes by
65d03bae8b but never found its
way into the official v15.2.0 release notes.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
The release note about bucket notifications being encapsulated in a JSON
array was added to PendingReleaseNotes by
19832a0dae but it never found its way into
the actual v15.2.0 release notes.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
The release note about OSDs self-classifying themselves as "nvme" in
certain rare cases was added to PendingReleaseNotes by
1d0fd5603e but it never found its way into
the official v15.2.0 release notes.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
The release note about restoring the original behavior of the rados
tool's "-o" option was added to PendingReleaseNotes by
d2b806070d but it never found its way into
the official v15.2.0 release notes.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
This note was added to PendingReleaseNotes by
0efa73f029 but never found its way into
the official v15.2.0 release notes.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
Adds option `mon_allow_pool_size_one` which will be disabled by default
to ensure pools are not configured without replicas.
If the user still wants to use pool size 1, they will have to change the
value of `mon_allow_pool_size_one` to true and then have to pass flag
`--yes-i-really-mean-it` to cli command:
Example:
`ceph osd pool test set size 1 --yes-i-really-mean-it`
Fixes: https://tracker.ceph.com/issues/44025
Signed-off-by: Deepika Upadhyay <dupadhya@redhat.com>
use APIs instead of apis to be consistent throughout.
fixes: https://tracker.ceph.com/issues/44374
Signed-off-by: Deepika Upadhyay <dupadhya@redhat.com>
Before this, "mds_join_fs" config enforced a preference for the standby
the monitors would select. Now the monitors actively enforce this
by purposefully removing an MDS wither lower "affinity". An MDS standby
has highest affinity if its mds_join_fs is the file system in question
or a vanilla standby (no mds_join_fs).
Fixes: https://tracker.ceph.com/issues/43392
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
1GB is too low as a default and usually results in cache size warnings
at that size; the MDS will struggle to maintain such a small cache size
for most workloads.
Fixes: https://tracker.ceph.com/issues/43182
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
Remove last bits of support for 'mds_cache_size'.
'mds_cache_memory_limit' is preferred.
Fixes: https://tracker.ceph.com/issues/41951
Signed-off-by: Ramana Raja <rraja@redhat.com>
osd/OSDMap: Show health warning if a pool is configured with size 1
Reviewed-by: Sage Weil <sweil@redhat.com>
Reviewed-by: David Zafman <dzafman@redhat.com>
Reviewed-by: Neha Ojha <nojha@redhat.com>
0b369e1aff masked the original behaviour of '-o' which was to indicate
'outfile' as documented in the man page. Changing object-size to capital
o will restore the original behaviour.
Fixes: https://tracker.ceph.com/issues/42477
Signed-off-by: Brad Hubbard <bhubbard@redhat.com>
Introduce a config option called 'mon_warn_on_pool_no_redundancy' that is
used to show a health warning if any pool in the ceph cluster is
configured with a size of 1. The user can mute/unmute the warning using
'ceph health mute/unmute POOL_NO_REDUNDANCY'.
Add standalone test to verify warning on setting pool size=1. Set the
associated warning to 'false' in ceph.conf.template under qa/tasks so
that existing tests do not break.
Fixes: https://tracker.ceph.com/issues/41666
Signed-off-by: Sridhar Seshasayee <sseshasa@redhat.com>
Currently, if a client requests the 1000 next entries from a bucket,
each bucket index shard will receive a request for the 1000 next
entries. When there are hundreds, thousands, or tens of thousands of
bucket index shards, this results in a huge amplification of the
request, even though only 1000 entries will be returned.
These changes reduce the per-bucket index shard requests. These also
allow re-requests in edge cases where all of one shard's returned
entries are consumed. Finally these changes improve the determination
of whether the resulting list is truncated.
Signed-off-by: J. Eric Ivancich <ivancich@redhat.com>