`rest/test-restful.sh` calls `test_mgr_rest_api.py`, which in turn
calls
```
('patch', '/config/osd', {'pause': True}),
```
and rest module translates it to `ceph osd set key=pause`
Signed-off-by: Kefu Chai <kchai@redhat.com>
* refs/pull/29034/head:
doc/mgr/crash: document missing commands, options
qa/suites/rados/singleton/all/test-crash: whitelist RECENT_CRASH
qa/suites/rados/mgr/tasks/insights: whitelist RECENT_CRASH
qa/tasks/mgr/test_insights: crash module now rejects bad crash reports
mgr/telemetry: fix remote into crash do_ls()
mgr/crash: don't make these methods static
mgr/BaseMgrModule: handle unicode health detail strings
mgr/crash: verify timestamp is valid
qa/suites/mgr: whitelist RECENT_CRASH
mgr/crash: remove unused var
mgr/crash: remove unused import 'six'
qa/workunits/rados/test_crash: health check
mgr/crash: improve validation on post
mgr/crash: automatically prune old crashes after a year
mgr/crash: raise RECENT_CRASH warning for recent (new) crashes
mgr/crash: add 'crash ls-new'
mgr/crash: add option and serve infra
mgr/crash: keep copy of crashes in memory
mgr/pg_autoscaler: adjust style to match built-in tables
mgr/crash: make 'crash ls' a nice table with a NEW column
mgr/crash: nicely format 'crash info' output
mgr/crash: add 'crash archive <id>', 'crash archive-all' commands
Reviewed-by: Neha Ojha <nojha@redhat.com>
The mgr's libcephfs client gets evicted after the mgr fails over.
Whitelist the message.
Fixes: http://tracker.ceph.com/issues/40867
Signed-off-by: Sage Weil <sage@redhat.com>
This warning is caused by the recent changes to the volumes
module that cache the CephFS handles.
Commit 5c41e949af9acabd612b0644de0603e374b4b42a
Signed-off-by: Ricardo Dias <rdias@suse.com>
When we are doing cache tiering, we are more sensitive to short PG logs
because the dup op entries are not perfectly promoted from the base to
the cache.
See:
http://tracker.ceph.com/issues/38358http://tracker.ceph.com/issues/24320
This works around the problem by not testing short pg logs in combination
with cache tiering. This works because the short_pg_log.yaml fragment
sets the short log in the [global] section but the cache workloads overload
it (back to a large/default value) in the [osd] section.
Signed-off-by: Sage Weil <sage@redhat.com>
If the leader is the one with the accurate clock, it can still
form quorum, but if the leader has the skewed clock, all other mons appear
skewed from its perspective and no quorum is formed. This leads to
intermittent failures, depending on the non-deterministic teuthology
deployment order and how the mon IPs sort.
Fix by reducing the skew. This is enough skew to trigger a warning, but
not enough that it will break quorum. This ensures that the parts of the
teuthology test that issue random mon commands won't fail (e.g., 'ceph osd
dump').
Fixes: http://tracker.ceph.com/issues/40112
Signed-off-by: Sage Weil <sage@redhat.com>
- don't specify ceph.py options in the ceph.py
- instead, specify them in the per-version facet
Note that we don't currently have a way to do v2 only for the mon IPs, so
in the v2only cases, we are still binding the mons to v1.
Signed-off-by: Sage Weil <sage@redhat.com>
in this change,
* suites/rados/upgrade: luminous-x-singleton => mimic-x-singleton
* suites/upgrade: luminous-x => nautilus-x
we support upgrade from n to n+2 release. otherwise monitor refuses to
do so:
mon.a@-1(probing) e1 current monmap has min_mon_release 15 (luminous)
which is >2 releases older than me 15 (octopus), stopping.
Fixes: https://tracker.ceph.com/issues/38845
Signed-off-by: Kefu Chai <kchai@redhat.com>
There is a lot of good stuff going on here, but nobody is investing in xio
and it is not expected to be the path forward for RDMA. If that ever
changes, we can resurrect the code. Until then, let's clean up the tree
and reduce friction for changes going forward.
Signed-off-by: Sage Weil <sage@redhat.com>
Hammer doesn't have async, and nautilus doesn't have simple. Expand the
async default (v2v1) and v1-only choice and the client choice.
Signed-off-by: Sage Weil <sage@redhat.com>
We've disabled the "clean" shutdown in ceph-mgr due to
https://tracker.ceph.com/issues/38621
Until then, no valgrind leak checks!
Signed-off-by: Sage Weil <sage@redhat.com>