Patrick Donnelly
a84e3c89bf
qa: thrash max_mds and deactivate ranks
...
Fixes: http://tracker.ceph.com/issues/10792
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
2017-07-06 22:29:41 -07:00
Patrick Donnelly
5b87301192
Merge remote-tracking branch 'upstream/pull/15822/head' into master
...
* upstream/pull/15822/head:
qa: add timeout/repeat to pool df
Reviewed-by: John Spray <jspray@redhat.com>
2017-07-06 22:14:32 -07:00
Patrick Donnelly
77b707a6ea
Merge remote-tracking branch 'upstream/pull/15950/head' into master
...
* upstream/pull/15950/head:
qa: update thrash max mds testing
Reviewed-by: Patrick Donnelly <pdonnell@redhat.com>
2017-07-06 22:11:18 -07:00
Patrick Donnelly
97cdb1e34a
Merge remote-tracking branch 'upstream/pull/15817/head' into master
...
* upstream/pull/15817/head:
qa: wait for healthy cluster before testing pins
Reviewed-by: Zheng Yan <zyan@redhat.com>
2017-07-06 21:36:34 -07:00
Patrick Donnelly
2cb42a4dbf
Merge remote-tracking branch 'upstream/pull/13770/head' into master
...
* upstream/pull/13770/head:
tasks/cephfs: add TestStrays.test_replicated_delete_speed
Reviewed-by: Patrick Donnelly <pdonnell@redhat.com>
2017-07-06 21:33:03 -07:00
Sage Weil
d02c9a5b01
Merge pull request #16122 from tchaikov/wip-qa-warning-msg
...
qa/tasks/ceph: should be "Waiting for all PGs", not "all osds"
Reviewed-by: xie xingguo <xie.xingguo@zte.com.cn>
2017-07-05 10:26:56 -05:00
Mykola Golub
2a9f56f818
Merge pull request #15860 from dillaman/wip-20168
...
librbd: fail IO request when exclusive lock cannot be obtained
Reviewed-by: Mykola Golub <mgolub@mirantis.com>
2017-07-05 14:52:55 +03:00
Kefu Chai
2cae544cf2
qa/tasks/ceph: should be "Waiting for all PGs", not "all osds"
...
amend the warning message
Signed-off-by: Kefu Chai <kchai@redhat.com>
2017-07-05 18:58:58 +08:00
Mykola Golub
866cf72440
Merge pull request #15956 from dillaman/wip-librbd-devstack
...
test: fix failing rbd devstack teuthology test
Reviewed-by: Mykola Golub <mgolub@mirantis.com>
2017-07-01 15:02:29 +03:00
Josh Durgin
4bcd6f666b
Merge pull request #14760 from dzafman/wip-19657
...
osd: On EIO from read recover the primary replica from another copy
Reviewed-by: Josh Durgin <jdurgin@redhat.com>
2017-06-29 14:00:18 -07:00
John Spray
623f1240a2
tasks/cephfs: add TestStrays.test_replicated_delete_speed
...
Reproducer for http://tracker.ceph.com/issues/16914
Signed-off-by: John Spray <john.spray@redhat.com>
2017-06-29 17:21:57 +01:00
Sage Weil
5712698787
Merge pull request #15894 from liewegas/wip-no-default-pool
...
mon,osd: do not create rbd pool by default
2017-06-29 11:21:42 -05:00
Yuri Weinstein
ea87f42cac
Merge pull request #15913 from smithfarm/wip-20417
...
tests: rgw.py: put client roles in a separate list
Reviewed-by: Casey Bodley <cbodley@redhat.com>
2017-06-28 08:07:54 -07:00
Sage Weil
93918b0be5
test,qa/workunits: fix a zillion tests
...
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-28 10:52:49 -04:00
Sage Weil
fc486c6fbd
qa/tasks/ceph: create rbd pool after starting the osds
...
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-28 10:52:49 -04:00
Sage Weil
fc7afc239f
Merge pull request #15853 from liewegas/wip-simpler-ceph
...
qa/tasks/ceph: simplify ceph deployment slightly
Reviewed-by: Kefu Chai <kchai@redhat.com>
2017-06-27 14:13:53 -05:00
Jason Dillaman
fdc4c9221f
qa/tasks/qemu: rbd cache is enabled by default
...
Signed-off-by: Jason Dillaman <dillaman@redhat.com>
2017-06-27 14:03:34 -04:00
Sage Weil
e7006d06fb
qa/tasks/ceph: explicitly add osds to crush map for upgrades
...
Before kraken, ceph-osd didn't add itself to crush... ceph-osd-prestart.sh
did it. And ceph.py doesn't use that.
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-27 12:52:35 -04:00
Jason Dillaman
6d4b32638c
qa/task/qemu: ensure test has access to all generated log files
...
Signed-off-by: Jason Dillaman <dillaman@redhat.com>
2017-06-27 12:48:11 -04:00
Jason Dillaman
f4f26b86d5
qa/suites/rbd: ensure guest VM updates to Ceph version under test
...
Signed-off-by: Jason Dillaman <dillaman@redhat.com>
2017-06-27 12:44:49 -04:00
Sage Weil
80e2b156e0
qa/tasks/dump_stuck: fix for active+clean+remapped
...
In d24a888665
we made remapped a clean
state but didn't fix this test.
Fixes: http://tracker.ceph.com/issues/20431
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-27 12:01:07 -04:00
Yan, Zheng
8d1828dc60
qa: update thrash max mds testing
...
Current monitor only allows deactivating one mds at a time. Besides,
the mds to deactivate should have max rank id.
Signed-off-by: "Yan, Zheng" <zyan@redhat.com>
2017-06-27 22:08:26 +08:00
Patrick Donnelly
95c0ca6a2b
qa: add timeout/repeat to pool df
...
Fixes: http://tracker.ceph.com/issues/20212
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
2017-06-27 06:50:15 -07:00
Nathan Cutler
9778643c67
tests: rgw.py: put client roles in a separate list
...
Fixes: http://tracker.ceph.com/issues/20417
Signed-off-by: Nathan Cutler <ncutler@suse.com>
2017-06-27 14:17:56 +02:00
Sage Weil
dfa95db57d
qa/tasks/radosbench: increase timeout
...
The current timeout isn't enough in some cases (powercycle thrashing leaves
osds down for a long time because rebooting is so slow).
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-26 15:09:19 -04:00
Zack Cerza
5442b5ff25
Merge pull request #15859 from smithfarm/wip-swift-task-move-master
...
tests: move swift.py task from teuthology to ceph, phase one (master)
2017-06-26 10:16:42 -06:00
Nathan Cutler
2bc0c7c617
tests: swift.py: tweak imports
...
The ".." form only works within the teuthology repo. With swift.py now in the
Ceph repo, we have to be explicit.
Error message was: "ValueError: Attempted relative import beyond toplevel
package
Signed-off-by: Nathan Cutler <ncutler@suse.com>
2017-06-25 12:42:36 +02:00
Casey Bodley
6b42352b70
Merge pull request #14624 from ceph/wip-s3a-hadoop
...
qa/tasks: S3A hadoop task to test s3a with Ceph
Reviewed-by: Casey Bodley <cbodley@redhat.com>
2017-06-23 13:46:05 -04:00
John Spray
b6cfa35458
qa: no longer need to explicitly enable multimds
...
Signed-off-by: John Spray <john.spray@redhat.com>
2017-06-23 17:07:34 +01:00
David Zafman
33edfe3a0f
test: Add two new singleton test yamls radom-eio and thrash-eio
...
New option "random_eio" to Thrasher, sets 1 osd random read percentage
New option "objectsize" to radosbench task (-o bench option)
New option "type" to radosbench specify write, seq or rand
Signed-off-by: David Zafman <dzafman@redhat.com>
2017-06-23 08:09:15 -07:00
Nathan Cutler
aab3920977
Merge branch 'master' of /home/smithfarm/src/ceph/upstream/teuthology into wip-swift-task-move-master
2017-06-23 08:30:38 +02:00
Nathan Cutler
7b58ac97e9
tests: move swift.py task to qa/tasks
...
In preparation for moving this task from ceph/teuthology.git into ceph/ceph.git
The move is necessary because jewel-specific changes are needed, yet teuthology
does not maintain a separate branch for jewel. Also, swift.py is a
Ceph-specific task so it makes more sense to have it in Ceph.
Signed-off-by: Nathan Cutler <ncutler@suse.com>
2017-06-23 08:27:42 +02:00
Sage Weil
2c8904506a
qa/tasks/ceph: create osds before starting
...
This ensure they are in the OSDMap before the CRUSH
map, which avoids the OSD_ORPHAN health warning.
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-22 13:45:27 -04:00
Sage Weil
0eed85515a
qa/tasks/ceph: do not pregenerate osdmap
...
This is useless, and further leads to a situation
where we have a populated CRUSH map before any OSDs,
triggering the OSD_ORPHAN error.
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-22 13:45:26 -04:00
John Spray
38dccd2c72
Merge pull request #15548 from ukernel/wip-20196
...
mds: improvements for stray reintegration
Reviewed-by: John Spray <john.spray@redhat.com>
2017-06-22 06:46:27 -04:00
Orit Wasserman
c320fbd9f8
Merge pull request #15753 from pritha-srivastava/wip-rgw-s3tests-conf
...
rgw: Changes for s3test config file, to add user under a tenant.
Reviewed-by: Casey Bodely <cbodley@redhat.com>
Reviewed-by: Orit Wasserman <owasserm@redhat.com>
2017-06-22 11:00:26 +03:00
Patrick Donnelly
d4870a093c
qa: wait for healthy cluster before testing pins
...
Fixes: http://tracker.ceph.com/issues/20318
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
2017-06-21 13:21:32 -07:00
Vasu Kulkarni
14b6267cba
s3a task to test radosgw compatibility with hadoop s3a interface
...
Signed-off-by: Vasu Kulkarni <vasu@redhat.com>
2017-06-21 11:52:10 -07:00
Sage Weil
6a00ba0e26
qa/tasks/ceph_manager: get osds all in after thrashing
...
Otherwise we might end up with some PGs remapped, which means they won't
get scrubbed.
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-20 12:07:25 -04:00
Yan, Zheng
57e82edc9c
qa/cephfs: use ceph.dir.pin to trigger migration
...
Signed-off-by: "Yan, Zheng" <zyan@redhat.com>
2017-06-20 17:39:46 +08:00
Pritha Srivastava
5e94a9852c
rgw: Changes for s3test config file, to add user under a tenant.
...
Signed-off-by: Pritha Srivastava <prsrivas@redhat.com>
2017-06-20 12:57:24 +05:30
Sage Weil
04969eff23
qa/tasks/resolve_stuck_peering: start osd at end
...
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-19 14:28:28 -04:00
Sage Weil
cc902a1f6b
qa/tasks/ceph: osd_scrub_pgs: reissue scrub requests in loop
...
The scrub commands are not reliable: if the OSD doesn't happen to
be connected at the time the command is issued it may not get
delivered. Re-request scrubs for each PG that has not yet been
scrubbed so that we don't wait forever when the original request
is dropped.
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-19 12:00:12 -04:00
Sage Weil
32361a798f
qa/tasks/ceph: osd_scrub_pgs: tolerate down osd at initial scrub time
...
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-19 12:00:12 -04:00
Sage Weil
bdf40c546d
Merge pull request #15717 from liewegas/wip-20326
...
qa/tasks/ceph.py: tolerate active+clean+something
Reviewed-by: Kefu Chai <kchai@redhat.com>
2017-06-16 16:12:20 -05:00
Sage Weil
1565b86dc0
qa/tasks/ceph.py: tolerate active+clean+something
...
where something is, say, snaptrim. or maybe scrubbing.
or whatever.
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-15 22:29:28 -04:00
Sage Weil
f870cc5f28
qa/tasks/thrashosds: wait before wait_for_recovery
...
Make sure OSDs are up *and* they have flushed their PG stats before
waiting for recovery to ensure that we do not see a stale 'clean' state.
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-15 12:14:24 -04:00
Sage Weil
200abcee6d
qa/tasks/ceph: raise exception if scrubs time out
...
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-15 11:23:18 -04:00
Sage Weil
0d80c88667
qa/tasks/ceph: raise an exception if pgs are not clean
...
If this happens the preceding test should have cleaned
up (e.g., ceph.healthy:).
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-15 11:23:18 -04:00
Sage Weil
6fa9d32407
qa/tasks/ceph: osd_scrub_pgs: try a bit longer
...
I just saw a test fail that was still waiting for
scrubs to complete.
Signed-off-by: Sage Weil <sage@redhat.com>
2017-06-15 11:23:18 -04:00