Commit Graph

41 Commits

Author SHA1 Message Date
Kefu Chai
ac1e13ce63 qa/tasks/ceph: add "create_rbd_pool" option
True by default, but we can opt not to create the "rbd" pool.

Signed-off-by: Kefu Chai <kchai@redhat.com>
2017-10-17 11:15:32 +08:00
Douglas Fuller
5fafc03cb9 qa/cephfs: Allow deferred fs creation
Permit Filesystem objects to be created and settings modified before
calling Filesystem.create().

Signed-off-by: Douglas Fuller <dfuller@redhat.com>
2017-08-30 09:02:44 -04:00
Sage Weil
9da7e63c64 qa/tasks/ceph.py: tolerate flush pg stats exception
If the OSD doesn't see IO, it won't flush more pg/osd stats when the
luminous flag is not yet set (legacy pgmonitor mode).

Signed-off-by: Sage Weil <sage@redhat.com>
2017-08-08 12:08:31 -04:00
Sage Weil
d9a0145f8f Merge pull request #16824 from liewegas/wip-more-scrub-time
qa/tasks/ceph: wait longer for scrub
2017-08-05 13:35:55 -05:00
Sage Weil
62482ce82c qa/tasks/ceph: debug osd setup
I've seen a couple rbd runs that seem to skip the next block :/

Signed-off-by: Sage Weil <sage@redhat.com>
2017-08-05 13:53:26 -04:00
Sage Weil
d09606619f qa/tasks/ceph: wait longer for scrub
Signed-off-by: Sage Weil <sage@redhat.com>
2017-08-04 12:06:27 -04:00
Sage Weil
f681d6e794 Merge pull request #16794 from liewegas/wip-rbd-app
qa/tasks/ceph: enable rbd on rbd pool
2017-08-04 08:40:45 -05:00
Sage Weil
7ed9d1e8aa qa/tasks/ceph: enable rbd on rbd pool
Signed-off-by: Sage Weil <sage@redhat.com>
2017-08-04 09:40:26 -04:00
Josh Durgin
63693779fc qa: timeout when waiting for mgr to be available
Otherwise during upgrades we wait forever.

Signed-off-by: Josh Durgin <jdurgin@redhat.com>
2017-08-02 02:18:28 -04:00
Sage Weil
a40d94b163 qa/tasks/ceph: wait for pg stats to flush in healthy check
Signed-off-by: Sage Weil <sage@redhat.com>
2017-07-27 12:10:27 -04:00
Sage Weil
dfc9c36606 fix ceph.py 2017-07-17 16:27:13 -04:00
Sage Weil
c66da972df qa/tasks/ceph.py: create osds in order
We aren't passing id to legacy 'osd create', which means we have to go
in order!

Signed-off-by: Sage Weil <sage@redhat.com>
2017-07-17 15:22:47 -04:00
Sage Weil
fd39700d55 qa/tasks/ceph: wait for osds to come up before creating pool
Avoid health warnings.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-07-12 12:52:03 -04:00
Sage Weil
83aaa75991 qa/tasks/ceph: stop logging health on shutdown
Don't log health during actual teardown or we'll see
various scary messages unrelated to our test run.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-07-12 12:52:02 -04:00
Sage Weil
b5f81e7313 qa/tasks/ceph.py: no osd id to 'osd create' command
This isn't recognized by hammer, and we don't need it for jewel.

Fixes: http://tracker.ceph.com/issues/20548
Signed-off-by: Sage Weil <sage@redhat.com>
2017-07-10 10:43:16 -04:00
Jason Dillaman
a06f8844c7 qa/tasks/ceph: supply cluster name to ceph CLI
Signed-off-by: Jason Dillaman <dillaman@redhat.com>
2017-07-07 11:21:51 -04:00
Jason Dillaman
5a8967cde7 qa/tasks/ceph: don't hard-code cluster name when copying fsid
Signed-off-by: Jason Dillaman <dillaman@redhat.com>
2017-07-07 10:44:20 -04:00
Kefu Chai
2cae544cf2 qa/tasks/ceph: should be "Waiting for all PGs", not "all osds"
amend the warning message

Signed-off-by: Kefu Chai <kchai@redhat.com>
2017-07-05 18:58:58 +08:00
Sage Weil
fc486c6fbd qa/tasks/ceph: create rbd pool after starting the osds
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-28 10:52:49 -04:00
Sage Weil
fc7afc239f Merge pull request #15853 from liewegas/wip-simpler-ceph
qa/tasks/ceph: simplify ceph deployment slightly

Reviewed-by: Kefu Chai <kchai@redhat.com>
2017-06-27 14:13:53 -05:00
Sage Weil
e7006d06fb qa/tasks/ceph: explicitly add osds to crush map for upgrades
Before kraken, ceph-osd didn't add itself to crush... ceph-osd-prestart.sh
did it.  And ceph.py doesn't use that.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-27 12:52:35 -04:00
John Spray
b6cfa35458 qa: no longer need to explicitly enable multimds
Signed-off-by: John Spray <john.spray@redhat.com>
2017-06-23 17:07:34 +01:00
Sage Weil
2c8904506a qa/tasks/ceph: create osds before starting
This ensure they are in the OSDMap before the CRUSH
map, which avoids the OSD_ORPHAN health warning.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-22 13:45:27 -04:00
Sage Weil
0eed85515a qa/tasks/ceph: do not pregenerate osdmap
This is useless, and further leads to a situation
where we have a populated CRUSH map before any OSDs,
triggering the OSD_ORPHAN error.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-22 13:45:26 -04:00
Sage Weil
cc902a1f6b qa/tasks/ceph: osd_scrub_pgs: reissue scrub requests in loop
The scrub commands are not reliable: if the OSD doesn't happen to
be connected at the time the command is issued it may not get
delivered.  Re-request scrubs for each PG that has not yet been
scrubbed so that we don't wait forever when the original request
is dropped.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-19 12:00:12 -04:00
Sage Weil
32361a798f qa/tasks/ceph: osd_scrub_pgs: tolerate down osd at initial scrub time
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-19 12:00:12 -04:00
Sage Weil
bdf40c546d Merge pull request #15717 from liewegas/wip-20326
qa/tasks/ceph.py: tolerate active+clean+something

Reviewed-by: Kefu Chai <kchai@redhat.com>
2017-06-16 16:12:20 -05:00
Sage Weil
1565b86dc0 qa/tasks/ceph.py: tolerate active+clean+something
where something is, say, snaptrim.  or maybe scrubbing.
or whatever.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-15 22:29:28 -04:00
Sage Weil
200abcee6d qa/tasks/ceph: raise exception if scrubs time out
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-15 11:23:18 -04:00
Sage Weil
0d80c88667 qa/tasks/ceph: raise an exception if pgs are not clean
If this happens the preceding test should have cleaned
up (e.g., ceph.healthy:).

Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-15 11:23:18 -04:00
Sage Weil
6fa9d32407 qa/tasks/ceph: osd_scrub_pgs: try a bit longer
I just saw a test fail that was still waiting for
scrubs to complete.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-15 11:23:18 -04:00
Sage Weil
fa0b2164ad qa/tasks/ceph.py: add 'skip_mgr_daemons' option
For upgrades

Signed-off-by: Sage Weil <sage@redhat.com>
2017-03-29 11:39:26 -04:00
Sage Weil
7edca203d8 qa/tasks/ceph.py: give everyone mgr caps
Signed-off-by: Sage Weil <sage@redhat.com>
2017-03-29 11:39:26 -04:00
John Spray
73100305e5 Merge pull request #13262 from batrick/multimds-thrasher
Add multimds:thrash sub-suite and fix bugs in thrasher for multimds

Reviewed-by: John Spray <john.spray@redhat.com>
2017-03-07 14:29:18 +00:00
Sage Weil
f777d849e7 qa/tasks/ceph.py: debug which pgs aren't scrubbing
Signed-off-by: Sage Weil <sage@redhat.com>
2017-02-24 23:07:34 -05:00
John Spray
de5249436c Merge pull request #13359 from jcsp/wip-logrotate-sshexception
qa: handle SSHException in logrotate

Reviewed-by: Kefu Chai <kchai@redhat.com>
2017-02-22 10:05:07 +00:00
Sage Weil
86c0d07e32 qa/tasks/ceph.py: fix timing of wait-for-* and osd markdown
Mark down osds, *then* wait for them to come up or for the cluster to be
healthy!

Signed-off-by: Sage Weil <sage@redhat.com>
2017-02-18 21:12:23 -05:00
John Spray
6f9e11f03d qa: handle SSHException in logrotate
Yet another different type of exception we may get when
orchestra.run can't talk to a remote host.

Signed-off-by: John Spray <john.spray@redhat.com>
2017-02-10 17:16:24 +00:00
Patrick Donnelly
cf9e0da078
qa: use fs methods for setting configs
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
2017-02-06 14:07:13 -05:00
Patrick Donnelly
0098873fb7
qa: remove old comment
Filesystem is now cluster aware.

Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
2017-02-06 14:07:13 -05:00
Sage Weil
c01f2ee0e2 move ceph-qa-suite dirs into qa/ 2016-12-14 11:29:55 -06:00