Commit Graph

71 Commits

Author SHA1 Message Date
Sage Weil
3118d9a154 osd: replace require_*_osds flags with require_osd_release field
- OSDMap encode and decode translate between the flags and int
representations.
- OSDMap::Incremental only does decode; we do not expect to ever encode
an incremental osdmap for an old osd that sets any of these flags.
- the 'osd set' command still lets you set the jewel and kraken flags,
but not luminous.
- OSDMap::apply_incremental handles the conversion of legacy require flags
to the new field if the jewel or kraken flags have to be set before
starting the osd upgrade.
- clear out the legacy flags when we make the luminous transition only;
until then we keep using the old flag in the encoded and decoded version
(although the require_osd_release field will be accurate in memory in all
cases).

Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-29 21:33:17 -04:00
Sage Weil
9a22b1a6ac qa/suites/rados/thrash-luminous: redirects require luminous
...so skip the at-end variant of d-luminous.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-26 22:52:48 -04:00
Sage Weil
a4247dd594 Merge branch 'wip-extensible_tier-redirect' of git://github.com/myoungwon/ceph into wip-redirect 2017-05-26 22:50:14 -04:00
Sage Weil
d2dac535f5 qa/suites/rados/singleton-nomsgr/health...: whitelist wrongly marked me down
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-24 17:58:12 -04:00
Sage Weil
85e2f3f31d qa/suites/rados/singleton-nomsgr: fix syntax
This parsed out as

  tasks:
  - install: null
  - ceph:
      conf:
        osd: osd max object name len = 400 osd max object namespace len = 64
  - workunit:
      clients:
        all:
        - rados/test_health_warnings.sh

which is clearly not correct.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-24 16:46:00 -04:00
Sage Weil
10a88f1c62 Merge pull request #15245 from liewegas/wip-20058
qa/suites/rados/*/at-end: wait for healthy before scrubbing
2017-05-23 21:47:13 -05:00
Sage Weil
88776b9c74 Merge pull request #15202 from liewegas/wip-19935
qa/suites/rados: temporarily remove scrub_test from basic/ until post-luminous
2017-05-23 15:37:08 -05:00
Sage Weil
ce654c5133 qa/suites/rados/*/at-end: wait for healthy before scrubbing
The scrub_pgs command also waits for healthy for a while, but fails
silently if it times out, which means the subsequent scrubs will also
fail to clean up.

This forces an earlier failure that does not obscure the root cause.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-23 14:12:24 -04:00
Sage Weil
657453d349 qa/suites/rados/singleton-nomsg/health-warnings: behave on ext4
We may land on an ext4 root partition.

Fixes: http://tracker.ceph.com/issues/20043
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-22 12:53:51 -04:00
Sage Weil
740f17ebf9 qa/suites/rados: temporarily remove scrub_test from basic/ until post-luminous
We don't want to do the at-end.yaml scrubbing business with this test.
Move it into a separate collection until after luminous.

I have a todo item on the post-luminous cleanup list to avoid forgetting
to move this back.

Fixes: http://tracker.ceph.com/issues/19935
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-22 10:15:23 -04:00
Yuri Weinstein
ef1c02406d Merge pull request #14914 from neha-ojha/wip-mon-crushtype-health-warn
mon: add crush type down health warnings

Reviewed-by: Sage Weil <sage@redhat.com>
Reviewed-by: Josh Durgin <jdurgin@redhat.com>
2017-05-18 13:12:45 -07:00
Sage Weil
739cbc98c3 Merge pull request #15117 from liewegas/wip-luminous-legacy-check
qa/suites/rados/thrash: make sure osds have map before legacy scrub

Reviewed-by: Josh Durgin <jdurgin@redhat.com>
2017-05-18 13:28:14 -05:00
Neha Ojha
6ecb03299a mon: add test for crush type down health warnings
Signed-off-by: Neha Ojha <nojha@redhat.com>
2017-05-16 14:16:19 -07:00
myoungwon oh
a07ad9fe80 qa/suites/rados/thrash: add redirect test cases
Signed-off-by: Myoungwon Oh omwmw@sk.com
2017-05-17 05:47:12 +09:00
Sage Weil
e57ecb64f0 qa/suites/rados/thrash: make sure osds have map before legacy scrub
The OSDs must have a map reflecting the require_luminous flag in order
for the legacy conversion to happen.  A quick rados bench should ensure
that.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-16 12:28:40 -04:00
Sage Weil
80038f50b2 qa/suites/rados/threas-erasure-code-*: fix luminous facet
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-15 10:22:25 -04:00
Sage Weil
8fdbeb5bb9 qa/suites/rados/thrash-erasure-code-overwrites: fix
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-15 10:21:58 -04:00
Sage Weil
d0a73ec955 Merge pull request #13610 from liewegas/wip-snapset
osd: eliminate snapdir objects and move clone snaps vector into SnapSet

Reviewed-by: Josh Durgin <jdurgin@redhat.com>
2017-05-11 21:18:11 -05:00
Sage Weil
1de9c90776 qa/suites: set initial require_min_compat_client
For cases where we are selecting crush tunables beyond the default
min of hammer.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-09 11:32:56 -05:00
Sage Weil
8bd54abc74 qa/suites/rados: at end, scrub pgs, verify no legacy snapsets
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-05 13:39:14 -04:00
Sage Weil
5f5f370925 qa/suites/rados: switch require-luminous facet to use full_sequential_finally
This lets us run multiple cleanup steps right before ceph
teardown.

Note that we drop the facet from multimon/ because it
doesn't factor out cluster creation before this step
properly.  That's fine because the require_luminous
cleanup shouldn't be related to the multimon tests.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-05 13:39:14 -04:00
Sage Weil
83dcc988db qa/suites/rados/verify: refactor thrash and cluster create
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-05 13:39:14 -04:00
Sage Weil
c9b93af06a qa/suites/rados/monthrash: simplify
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-05 13:39:14 -04:00
Sage Weil
6e20af4fd1 qa/suites/rados/thrash-erasure-code: factor out cluster create
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-05 13:39:14 -04:00
Sage Weil
112e232054 qa/suites/rados/basic: factor out cluster start
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-05 13:39:14 -04:00
Sage Weil
dc39cbab18 qa/suites/rados/thrash: extra cluster create
Signed-off-by: Sage Weil <sage@redhat.com>
2017-05-05 13:39:14 -04:00
Yuri Weinstein
1cce4c7f68 Merge pull request #14723 from neha-ojha/objclass-sdk
osd: objclass sdk

Reviewed-by: Josh Durgin <jdurgin@redhat.com>
Reviewed-by: Sage Weil <sage@redhat.com>
Reviewed-by: Kefu Chai <kchai@redhat.com>
2017-05-03 09:28:59 -07:00
Yuri Weinstein
9cb79d2fe3 qa/added overrides
Signed-off-by: Yuri Weinstein <yweinste@redhat.com>
2017-05-02 15:06:49 -07:00
Yuri Weinstein
35171b97a3 Merge pull request #14687 from jcsp/wip-mgr-suite
qa/suites: move mgr tests into rados suite

Reviewed-by: Sage Weil <sage@redhat.com>
2017-05-02 09:06:51 -07:00
Neha Ojha
522bd84dc0 objclass-sdk: add override to whitelist the sdk class
Signed-off-by: Neha Ojha <nojha@redhat.com>
2017-05-01 16:28:54 -07:00
Kefu Chai
86c9137fa0 qa/suites/rados/singleton-bluestore: concat settings
otherwise the bluestore tests will fail with failures like

qa/workunits/cephtool/test.sh:1343: test_mon_osd_pool:  ceph osd pool set ec_test allow_ec_overwrites true
Error EINVAL: pool must only be stored on bluestore for scrubbing to work: osd.1 uses filestore
qa/workunits/cephtool/test.sh:1343: test_mon_osd_pool:  return 1

Signed-off-by: Kefu Chai <kchai@redhat.com>
2017-04-30 23:19:07 +08:00
Nathan Cutler
e69a9763a0 tests: rados: move cephtool.yaml to new singleton/bluestore subsuite
cephtool.yaml is bluestore-only, yet it was in singleton/ which runs against a
generalized objectstore matrix.

Fixes: http://tracker.ceph.com/issues/19797
Signed-off-by: Nathan Cutler <ncutler@suse.com>
2017-04-28 15:36:50 +02:00
Kefu Chai
6fa16c4477 Merge pull request #14584 from tchaikov/wip-19631
qa/suites: Revert "qa/suites: add mon-reweight-min-pgs-per-osd = 4"

Reviewed-by: Sage Weil <sage@redhat.com>
2017-04-21 22:56:21 +08:00
John Spray
01c2016aca qa/suites: add third mon to mgr test
There were always meant to be three, having two
was a typo.

Signed-off-by: John Spray <john.spray@redhat.com>
2017-04-21 11:27:50 +01:00
David Zafman
7d579ff12b cephtool.yaml: Add "failsafe" osd error messages to whitelist
Signed-off-by: David Zafman <dzafman@redhat.com>
2017-04-20 15:47:12 -07:00
John Spray
017f260c9b qa/suites: disable scrub on shutdown in mgr test
The tests that exercise mgr failover do not necessarily
leave a happy working mgr daemon in place, and since
pg dump moved into the mgr, that means they should
not try and call "pg dump" to validate PG state on shutdown.

Signed-off-by: John Spray <john.spray@redhat.com>
2017-04-20 22:34:42 +01:00
John Spray
0829093383 qa/suites: move mgr tests into rados suite
Signed-off-by: John Spray <john.spray@redhat.com>
2017-04-20 22:34:42 +01:00
Kefu Chai
b3075c485c qa/suites: Revert "qa/suites: add mon-reweight-min-pgs-per-osd = 4"
This reverts commit ef7583d717.

Signed-off-by: Kefu Chai <kchai@redhat.com>
2017-04-20 14:18:21 +08:00
Josh Durgin
a3e5cba3f1 qa: move ec overwrites tests to a bluestore-only subsuite
Signed-off-by: Josh Durgin <jdurgin@redhat.com>
2017-04-19 17:45:43 -07:00
Josh Durgin
6fba80c1fa osd, OSDMonitor, qa: mark ec overwrites non-experimental
Keep the pool flag around so we can distinguish between a pool that
should maintain hashes for each chunk, and a missing one is a bug, vs
an overwrites pool where we rely on bluestore checksums for detecting
corruption.

Signed-off-by: Josh Durgin <jdurgin@redhat.com>
2017-04-19 17:45:43 -07:00
Sage Weil
ee1bb01a54 Merge pull request #14556 from liewegas/wip-pgupmap
osd: pg-remap -> pg-upmap

Reviewed-by: David Zafman <dzafman@redhat.com>
2017-04-19 17:07:01 -05:00
Sage Weil
ce188e8fdf osd: pg-remap -> pg-upmap
'remap' is to non-specific a name.  In particular, it
sounds like it is related to the 'remapped' PG state
but in reality it is not related.

'upmap' or 'pg-upmap' is more specific: it maps a pgid
to the 'up' set value (or item)

Signed-off-by: Sage Weil <sage@redhat.com>
2017-04-18 12:59:40 -04:00
Sage Weil
d82ac99d70 Merge pull request #14490 from liewegas/wip-no-luminous
mon: add mon_debug_no_require_luminous

Reviewed-by: Kefu Chai <kchai@redhat.com>
2017-04-14 20:05:45 -05:00
Sage Weil
8c6820a433 qa/suites/rados/thrash: do not thrash cluster full without luminous flag
The interface changes with the luminous flag; this will fail.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-04-14 12:04:47 -04:00
Sage Weil
4857f51e68 qa/suites/rados: expand other collections with no-require-luminous
Signed-off-by: Sage Weil <sage@redhat.com>
2017-04-14 11:45:05 -04:00
Sage Weil
8fd450e986 qa/suites/rados: require_luminous_osds at end of run, then scrub
Signed-off-by: Sage Weil <sage@redhat.com>
2017-04-13 17:11:19 -04:00
Kefu Chai
ef7583d717 qa/suites: add mon-reweight-min-pgs-per-osd = 4
so we can run reweight commands in a small cluster with 3 OSD and 27
PGs.

Signed-off-by: Kefu Chai <kchai@redhat.com>
2017-04-12 23:29:10 +08:00
Sage Weil
14e7d6351a Merge pull request #14198 from liewegas/wip-fs
qa/suites: drop 'fs' facet, and add 'objectstore' facet where missing

Reviewed-by: Josh Durgin <jdurgin@redhat.com>
Reviewed-by: Jason Dillaman <dillaman@redhat.com>
2017-04-08 13:21:03 -05:00
Sage Weil
d403d42ac8 qa/suites/rados/upgrade: don't initially start mgr daemon
Signed-off-by: Sage Weil <sage@redhat.com>
2017-03-29 11:39:26 -04:00
Sage Weil
271a7588b5 qa/suites: run mgr daemon(s)
Everything up upgrade/, which will be slightly tricky.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-03-29 11:39:26 -04:00