Commit Graph

154 Commits

Author SHA1 Message Date
John Spray
e51cd6bee3 tasks/cephfs: prefix override in FuseMount
For asok and invokation of ceph CLI.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-10-02 17:41:16 +01:00
John Spray
7e6f8d4905 tasks/cephfs: use stdin/stdout in Filesystem._read_data_xattr
...so that it doesn't rely on a temp local file and using
run.Raw(">") in commands.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-10-02 17:41:16 +01:00
John Spray
48cf533f0a tasks/cephfs: handle endpoint not connected
Previously failure to stat mnt dir was interpreted
as being unmounted.  For "transport endpoint no connected"
error we do want to recognise that it is mounted, albeit
with no ceph-fuse process.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-10-02 17:41:16 +01:00
John Spray
120f9d882e tasks/cephfs: use Filesystem.rados instead of direct CLI
Because it's slicker and because I can wrap it for vstart.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-10-02 17:41:16 +01:00
John Spray
bab4c7aa71 tasks/cephfs: add Filesystem.exists
Use this during test setup to check whether
a filesystem is configured at all, before
trying to tear it down.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-10-02 17:41:16 +01:00
John Spray
e2f8b48d94 tasks/cephfs: add Filesystem._prefix
So that my vstart subclass can put ./ before
all the commands.

One could set $PATH, but I like to unambiguously point
it at the local built binaries in case someone also
has some systemwide packages.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-10-02 17:41:15 +01:00
John Spray
f58c119d90 tasks/cephfs: use raw_cluster_cmd instead of manual "ceph"
It's less repetitive, and I can override it for use with
vstart.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-10-02 17:41:15 +01:00
John Spray
99a8a123da tasks/cephfs: add Filesystem.delete_all
So that we can clear down filesysteems created by
someone other than ourselves (like vstart.sh)

Signed-off-by: John Spray <john.spray@redhat.com>
2015-10-02 17:41:15 +01:00
John Spray
ad562f4b79 tasks/cephfs: make Filesystem rados command overridable
Signed-off-by: John Spray <john.spray@redhat.com>
2015-10-02 17:41:15 +01:00
John Spray
7884d5b555 tasks/cephfs: use raw_cluster_cmd in Filesystem
...instead of manually invoking 'ceph' -- so that
it's easier to override this for vstart.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-10-02 17:41:15 +01:00
Gregory Farnum
6add6ba2df Merge pull request #461 from ceph/wip-client-oldest-tid
cephfs/test_client_limits.py: update test_client_oldest_tid

Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-10-01 22:37:15 -07:00
Sage Weil
a4f9bdbe54 tasks/cephfs: mds allow
Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-28 12:38:04 -04:00
Gregory Farnum
e3c99476b0 Merge pull request #453 from ceph/wip-11835
tasks/cephfs: time out on ceph-fuses that don't die

Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-09-21 16:09:48 -07:00
Greg Farnum
f97fde6c26 tasks/cephfs: switch unmount timeout to 15 minutes
Our ffsb and fsync tests contain so many small writes at random offsets
that it can take >10 minutes to commit all of them to disk if we get
a slower OSD cluster. 15 minutes is still a plenty-fast timeout for
this stage compared to just hanging and losing the logs!

Signed-off-by: Greg Farnum <gfarnum@redhat.com>
2015-09-16 14:27:14 -07:00
John Spray
8e275049a0 tasks/cephfs: fix mount_timeout config loading
Previously were defaulting to a string, which
always compared greater than elapsed, so never
timed out.

Fixes: #12820
Signed-off-by: John Spray <john.spray@redhat.com>
2015-09-01 11:31:23 +01:00
John Spray
ffe79e9f31 tasks/cephfs: timeout on wait in test_network_death
So that we can get a failure instead of a hang in

Signed-off-by: John Spray <john.spray@redhat.com>
2015-08-28 13:57:23 +01:00
John Spray
df8cecf479 tasks/cephfs: fix test_pool_perm teardown
Fixes: #12777
Signed-off-by: John Spray <john.spray@redhat.com>
2015-08-27 17:53:24 +01:00
John Spray
c1ca95cd16 tasks/cephfs: implement TestDataScan.test_stashed_layout
This is for testing how cephfs-data-scan handles
the new 'layout' xattr when found during recovery.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-08-13 14:29:35 +01:00
Gregory Farnum
961701e3ed Merge pull request #512 from ceph/wip-pool-perm-fixes
Fixes to test_pool_perm

Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-07-28 14:45:44 +01:00
John Spray
a13064bf21 Merge pull request #489 from dachary/wip-12320-ceph-fuse
tasks/ceph_fuse.py: virtual machines need flexible mount timeout

Reviewed-by: John Spray <john.spray@redhat.com>
2015-07-28 11:20:36 +01:00
John Spray
e33065f3c9 tasks/cephfs: reset auth caps in setUp
Signed-off-by: John Spray <john.spray@redhat.com>
2015-07-28 10:34:58 +01:00
John Spray
6864f882bc tasks/cephfs: don't hardcode data pool name in test_pool_perm
Signed-off-by: John Spray <john.spray@redhat.com>
2015-07-28 10:34:58 +01:00
John Spray
e3434b6d2f tasks/cephfs: test_pool_perm: reinstate proper caps on teardown
Signed-off-by: John Spray <john.spray@redhat.com>
2015-07-28 10:34:58 +01:00
John Spray
a0ccd69041 tasks/cephfs: don't hardcode client ID in test_pool_perm
Our teuthology yamls all happen to use "client.0" but
it's not guaranteed.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-07-28 10:34:58 +01:00
John Spray
36ec6f9e0b tasks/cephfs: test_pool_perm whitespace
PEP8-ize.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-07-28 10:34:57 +01:00
Loic Dachary
ec12f21a7e tasks/ceph_fuse.py: virtual machines need flexible mount timeout
1) add a wait time before the mount attempt to let the cluster get set up.
By default this should be skipped, but for VMs and known-slow systems we
can give them 60 seconds.
2) Make the timeout configurable, with a 30-second default, but override it
for VM tests.

http://tracker.ceph.com/issues/12320 Fixes: #12320

Signed-off-by: Loic Dachary <loic@dachary.org>
2015-07-14 13:03:01 +02:00
Gregory Farnum
ba6e8e6d94 Merge pull request #464 from ceph/wip-offline-backward
Tests for damage handling and cephfs-data-scan

Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-07-14 11:43:02 +01:00
Yan, Zheng
44dc0320fa Merge pull request #470 from ceph/wip-4161
Add cephfs TestBacktrace
2015-07-13 17:22:58 +08:00
Yan, Zheng
78a62d41b8 Merge pull request #466 from ceph/wip-damaged-fixes
Tests for wip-damaged-fixes
2015-07-13 17:03:14 +08:00
Yan, Zheng
835ad8b8f3 cephfs/test_auto_repair.py: flush journal after umount
When client capabilities get released, MDS may update corresponding
inodes' client writable range and mark those inodes dirty. The auto
repair test expects MDS to trim inodes from its cache, but MDS can't
trim dirty inodes. So we should flush journal after umount.

Fixes: #12172
Signed-off-by: Yan, Zheng <zyan@redhat.com>
2015-06-29 17:52:44 +08:00
John Spray
21abe9d928 tasks/cephfs: add test_backtrace
This is for verifying the new layout-writing behaviour.  While
we're at it, test that the pre-existing backtrace behaviours
are really happening (updating old_pools)

Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-25 17:19:03 +01:00
John Spray
fa16974063 tasks/cephfs: add TestDamage
Tests MDS response to badness in the metadata tool

Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-18 11:22:12 +01:00
John Spray
cc2e9ff579 tasks/cephfs: allow stdin for Filesystem.rados
Useful for puts when injecting data to test
corruption handling.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-18 11:22:12 +01:00
John Spray
655d197a6b tasks/cephfs: add CephFSMount.stat
Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-18 11:22:11 +01:00
John Spray
f7e932fb8f tasks/cephfs: add TestDamage
Tests MDS response to badness in the metadata tool

Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-17 14:22:17 +01:00
John Spray
40d0c05c35 tasks/cephfs: add TestDataScan
For testing the new cephfs-data-scan disaster recovery
utility.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-17 14:09:59 +01:00
John Spray
d2cec6a669 tasks/cephfs: add CephFSMount.ls
Just a helper for functions that inspect
the contents of the mount, such as in
test_data_scan

Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-17 14:06:23 +01:00
John Spray
e6afe66b24 tasks/cephfs: allow stdin for Filesystem.rados
Useful for puts when injecting data to test
corruption handling.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-17 14:06:23 +01:00
John Spray
718eaee5e5 tasks/cephfs: add CephFSMount.stat
Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-17 14:06:22 +01:00
John Spray
0f34d0785c tasks/cephfs: add Filesystem.data_scan wrapper
Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-17 14:06:22 +01:00
Yan, Zheng
786f854b52 cephfs/test_client_limits.py: update test_client_oldest_tid
the health warning message has been changed

Signed-off-by: Yan, Zheng <zyan@redhat.com>
2015-06-16 14:55:18 +08:00
John Spray
63a563d0e7 tasks/cephfs: fix race in test_full
Sometimes mount A would get a cap revoke when mount
B did its last IO, resulting in mount A's OSD epoch
getting updated too.

Fix by making sure mount B is the last one to have
done IO before we do the barrier, so that when
it does IO again after the barrier, mount A can't
be holding any caps that B would need.

Fixes: #11913
Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-11 11:21:26 +01:00
John Spray
07eb03acee tasks/cephfs: time out on ceph-fuses that don't die
For cases where we have e.g. poked the fuse abort
file for a process, but it's still not dying.  Because
this is a special class of error (unlike e.g. when
we force umount something because the network is gone)
raise the error instead of trying again to kill
the client.

Fixes: #11835
Signed-off-by: John Spray <john.spray@redhat.com>
2015-06-10 10:56:07 +01:00
Gregory Farnum
7f8b8660d9 Merge pull request #449 from ceph/wip-fsync
tasks/cephfs: add TestClientRecovery.test_fsync

Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-06-05 13:50:17 -07:00
John Spray
6d3b5c6b5e tasks/cephfs: add TestClientRecovery.test_fsync
To test that metadata written recently is
preserved across a client+server crash when
barriered with a directory fsync.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-05-29 14:33:01 +01:00
John Spray
df91f98e5b tasks/cephfs: fix typo in blacklist clearing
Broken in aa0ffb3

Signed-off-by: John Spray <john.spray@redhat.com>
2015-05-29 13:46:01 +01:00
Gregory Farnum
d9959b18ba Merge pull request #443 from ceph/wip-11301
tasks/cephfs: reset osd blacklist between tests

Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-05-27 14:29:26 -07:00
Gregory Farnum
4a79c75daf Merge pull request #447 from ceph/wip-11779
tasks/cephfs: fix timing in test_full

Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-05-27 14:17:27 -07:00
John Spray
5ff2743de4 tasks/cephfs: fix timing in test_full
The interval between writes was too short because
it was not taking account of the way OSDMap full
flags are set on tick rather than immediately.

Fixes: #11779

Signed-off-by: John Spray <john.spray@redhat.com>
2015-05-27 12:40:49 +01:00
Yan, Zheng
bd542353f5 tasks/cephfs: mount fusectl before listing fuse connections
Fixes: #11756
Signed-off-by: Yan, Zheng <zyan@redhat.com>
2015-05-26 22:18:57 +08:00
John Spray
aa0ffb313c tasks/cephfs: reset osd blacklist between tests
...to avoid OSDMap modifications happening in the
background due to blacklist expiry.

Fixes: #11301

Signed-off-by: John Spray <john.spray@redhat.com>
2015-05-22 11:04:46 +01:00
Gregory Farnum
fed8837c98 Merge pull request #442 from ceph/wip-client-cache-size
cephfs/test_client_limits.py: invalidate kernel dcache according to client cache size

Reviewed-by: John Spray <john.spray@redhat.com>
Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-05-15 10:36:16 -07:00
Gregory Farnum
8eb3255a4c Merge pull request #425 from ceph/wip-flock-optional
cephfs/mount: check FUSE version before invoking flock

Reviewed-by: John Spray <john.spray@redhat.com>
2015-05-14 21:52:30 -07:00
Yan, Zheng
c7e8e544a8 cephfs/test_client_limits.py: invalidate kernel dcache according to client_cache_size
Signed-off-by: Yan, Zheng <zyan@redhat.com>
2015-05-13 14:40:28 +08:00
Greg Farnum
f780f1894b cephfs/test_client_recovery: check FUSE version before invoking flock
flock only works properly on FUSE versions >=2.9, which is newer
than eg Ubuntu Precise. So check the version on our client mounts and
only test flock if it's at least that new.

Fixes: #9995

Signed-off-by: Greg Farnum <gfarnum@redhat.com>
2015-05-12 14:43:19 -07:00
John Spray
13c7f97d89 tasks/cephfs: cover pool quota in test_full
Run the same procedure as TestClusterFull, but
instead of limiting OSD memstore size, use pool
quota on the data pool.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-05-08 14:09:59 +01:00
Yan, Zheng
539117a38d tasks/cephfs: add test_pool_perm
This tests that client checks pool permission before
copy data to its cache.

Signed-off-by: Yan, Zheng <zyan@redhat.com>
2015-04-28 20:38:32 +08:00
Yan, Zheng
e850645fe4 cephfs/test_client_limits.py: test for client oldest tid warning
Signed-off-by: Yan, Zheng <zyan@redhat.com>
2015-04-21 23:29:47 +08:00
Greg Farnum
35b6fb436c Merge remote-tracking branch 'origin/wip-3645-fixup' into greg-testing-snapshot-rename 2015-04-19 18:21:08 -07:00
John Spray
f36011051c tasks/cephfs: cope with missing ctx.daemons attr
...so that we may be used with the ceph_deploy task
that doesn't set that up.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-19 17:46:18 -07:00
John Spray
8b61310ce3 tasks/cephfs: tweak use of mon for admin commands
... s/mon_remote/admin_remote/ and allow caller to pass
in which remote they want to use for that.  Enables use
with ceph_deploy task which does not give admin keys
to mons.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-19 17:46:18 -07:00
John Spray
ea7c39222a tasks/ceph: refactor legacy FS configuration check
Move up into Filesystem so that this can be used from
the ceph_deploy task as well.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-19 17:46:18 -07:00
John Spray
7bc8875b42 tasks: implement TestStrays.test_mv_hardlink_cleanup
Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-14 14:13:40 +01:00
John Spray
a2a8c21abe tasks: update test_journal_repair
This broke with recent Client changes that
do better caching of readdir results, such
that doing an ls twice is no longer sufficient
to see a fresh result after repair - we need
to remount instead.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-14 14:13:40 +01:00
John Spray
097ccbb5be tasks: update journal_repair test for 'damaged' state
To track recent change in master where instead of
crashing on missing MDSTable object we'll go
into damaged state.

Instead of catching a crash, handle the rank's
transition to the damanged state.  Leave the crash
handling code (unused for the moment) in the
Filesystem class in case it's needed elsewhere
soon.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-14 14:13:39 +01:00
John Spray
bd3ae1eeba tasks/cephfs: add test_strays
This tests the new purge file/ops throttling
in the MDS, via the new perf counters for
strays/purging.

Fixes: #10390
Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-14 14:13:39 +01:00
John Spray
abb635588f tasks/cephfs: add test_sessionmap
Tests for the persistence behaviour of SessionMap.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-14 14:13:39 +01:00
John Spray
2b5137bf06 tasks: generalise cephfs test runner
...to avoid having boilerplate in each test module,
and gain the ability to run them all in one go
with a nice test-by-test pass/fail report.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-14 14:13:39 +01:00
John Spray
79906e3d07 tasks/cephfs: better multiple-mds handling
Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-14 14:13:39 +01:00
John Spray
ce1196d62f tasks/cephfs: be tolerant of multiple MDSs
...as long as only one is active, all the ops
that default to talking to a single MDS should
be happy to talk to the active MDS, even if there
happens to be a standby lying around too.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-04-14 14:13:38 +01:00
John Spray
1647453d66 Merge pull request #356 from ceph/wip-10948
tasks/cephfs: don't run iptables in parallel

Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-03-03 20:25:09 +00:00
John Spray
5b2fd3d614 tasks/cephfs: don't run iptables in parallel
Where multiple MDSs were on the same node, trying
to concurrently update their firewall state was
causing an exception because the iptables command
errors out if another instance is already running.

Fixes: #10948
Signed-off-by: John Spray <john.spray@redhat.com>
2015-03-02 12:41:13 +00:00
John Spray
fc4e90930b tasks/cephfs: fix/improve fuse force umount (again)
teuthology helpfully escapes things for us so
the \; didn't need the backslash.  The logic
was still falling over in some cases too.

Additionally, make the FUSE /sys/ abort operation
more surgical by working out the connection name
of our own mount during mount().

Signed-off-by: John Spray <john.spray@redhat.com>
2015-02-26 23:08:33 +00:00
Alfredo Deza
0a65e90db2 remove uneeded variable assignment for call
Signed-off-by: Alfredo Deza <adeza@redhat.com>
2015-02-26 11:35:31 -05:00
Alfredo Deza
3a18cb29a8 remove redefined StringIO import
Signed-off-by: Alfredo Deza <adeza@redhat.com>
2015-02-26 11:35:07 -05:00
John Spray
03b0e1064c tasks/cephfs: fix fuse force unmount
This was broken in the case of multiple
mounts, and in the case of stuck mounts.

Fixes: #10702
Signed-off-by: John Spray <john.spray@redhat.com>
2015-02-10 14:58:23 +00:00
John Spray
bf13a4e062 tasks/mds_journal_repair: create new test
This tests the new #9883 repair functionality
where we selectively scrape dentries out of
the journal while the MDS is offline.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-01-19 19:36:34 +00:00
John Spray
997c0c3d62 tasks/cephfs: remove config from Filesystem cons
This was only used in get_first_mon, which doesn't actually
need the parameter itself.  Makes it easier to casually
use Filesystem from any place with a ctx to hand.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-01-09 10:09:53 +00:00
John Spray
4ff4819573 tasks/cephfs: umount unused clients before running
When unused clients were mounted during an fs new,
they would end up in a state where they stalled
on subsequent attempts to umount them (ceph-fuse
stalls on exit if it can't terminate its mds_session)

Signed-off-by: John Spray <john.spray@redhat.com>
2015-01-09 10:09:52 +00:00
John Spray
8630c4ab3b tasks/cephfs: move ROOT_INO into filesystem module
...so that other tests can get at it.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-01-09 10:09:52 +00:00
John Spray
5e3f5766d0 tasks/cephfs: be more precise in set_clients_block
Instead of blocking the whole port range (which
might make OSDs running on that node collateral
damage), read the MDS's port out of the MDS map
and just block that.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-01-09 10:09:52 +00:00
John Spray
cfbc277c1f tasks/cephfs: move common setUp/tearDown to parent
Now that we have more of these cases, there was lots
of duplication in setup and teardown.  For some tests
the "reset everything" setup/teardown is overkill,
but it's harmless.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-01-09 10:09:51 +00:00
John Spray
6d79e3918b tasks/cephfs: move wait_until helpers to testcase
...so that other test cases can use them

Signed-off-by: John Spray <john.spray@redhat.com>
2015-01-09 10:08:09 +00:00
John Spray
7696afd3fa tasks: create mds_flush task
This tests:
 * The new 'flush journal' asok command
 * That the resulting on disk structures are as expected
 * That cephfs-journal-tool is happy with the result

Fixes: #9881

Signed-off-by: John Spray <john.spray@redhat.com>
2014-12-19 13:40:53 +00:00
John Spray
849e259b96 tasks: add mds_full task
This is for exercising the CephFS behaviour
when the cluster becomes full.

Signed-off-by: John Spray <john.spray@redhat.com>
2014-12-18 12:45:31 +00:00
John Spray
f9609c4688 tasks/cephfs: let get_config use non-mon services
Previously was always using the default values of things
so querying mon instead of the appropriate service
worked fine.  However, for things we might want to
update on a per-test basis we need to go ask the
correct service what the setting really is.

Needed for osd_mon_report_interval_max in the ENOSPC
testing.

Signed-off-by: John Spray <john.spray@redhat.com>
2014-12-18 12:45:30 +00:00
John Spray
8fa6b154e3 tasks/cephfs: enable osd epoch get from mounts
Using new hooks for #7317

Signed-off-by: John Spray <john.spray@redhat.com>
2014-12-18 12:45:30 +00:00
John Spray
dedcc6483a tasks/cephfs: enable kclient for mds_* tasks
Signed-off-by: John Spray <john.spray@redhat.com>
2014-12-18 12:45:30 +00:00
John Spray
b4214f95f4 Merge pull request #257 from ceph/wip-fs-creation
Fix PG counts and refactor FS creation during setup

Reviewed-by: Loic Dachary <ldachary@redhat.com>
2014-12-17 11:37:09 +00:00
Yan, Zheng
09b82e2bd9 tasks: add test case for readonly MDS
Signed-off-by: Yan, Zheng <zyan@redhat.com>
2014-12-10 09:48:02 +08:00
John Spray
8fc3550002 tasks/ceph: move FS creation inside Filesystem
Leave the legacy handling out in cephfs_setup, move
the filesystem creation stuff into Filesystem.  I
anticipate this being the right place for it if/when
we have tests that want to do 'fs rm' 'fs new' type
cycles within themselves.

Signed-off-by: John Spray <john.spray@redhat.com>
2014-12-02 13:31:57 +00:00
John Spray
89beed1b0a tasks/ceph: conservative PG count in FS pools
This was tripping over the recent commit 42c85e80
in Ceph master, which tightens the limits on
acceptable PG counts per OSD, and was making
teuthology runs fail due to never going clean.

Rather than put in a new hardcoded count, infer
it from config.  Move some code around so that
the ceph task can get at a Filesystem object
to use in FS setup (this already has conf-getting
methods).

Signed-off-by: John Spray <john.spray@redhat.com>
2014-12-02 13:30:06 +00:00
Yan, Zheng
d761831133 tasks: add mds_auto_repair
New CephFS tests for MDS's auto repair functions. (So far the only
test case is verify/repair backtrace on fetch dirfrag)

Signed-off-by: Yan, Zheng <zyan@redhat.com>
2014-12-02 16:26:51 +08:00
Yan, Zheng
edb780a3c5 tasks/cephfs/mount: use seperate for testing flock and posix lock
Old version libfuse treats both flock and posix lock requests as posix
lock request. This is a workaround for the bug.

Fixes: #9995
Signed-off-by: Yan, Zheng <zyan@redhat.com>
2014-11-07 09:07:27 +08:00
Yan, Zheng
88133719b7 tasks/mds_client_recovery: file lock test
check that file lock doesn't get lost after an MDS restart

Signed-off-by: Yan, Zheng <zyan@redhat.com>
2014-10-10 16:54:29 +08:00
John Spray
6ac9efef9c tasks/cephfs: say which test failed in exception
Example:
Was: 'Test failure'
Now: Test failure: test_full_caps (tasks.mds_full.TestClusterFull)

Signed-off-by: John Spray <john.spray@redhat.com>
2014-10-08 16:27:44 +01:00
John Spray
0073e25d77 tasks: rename FuseMount.get_client_id to get_global_id
'client_id' was ambiguous because in other places it
meant the '0' in client.0, whereas here it means
the runtime-generated global ID of the client.

Signed-off-by: John Spray <john.spray@redhat.com>
2014-09-29 15:04:25 +01:00
John Spray
b77b3bec72 tasks: add mds_client_limits
New CephFS tests for the behaviour of the system while
enforcing its resource limits.

Signed-off-by: John Spray <john.spray@redhat.com>
2014-09-29 15:04:18 +01:00
John Spray
1fa15011a3 tasks: generalise CephFSTestCase
Some of this stuff could be even more general for embedding
unittest-style suites, but for the moment let's keep the cephfs
stuff in a walled garden.

Signed-off-by: John Spray <john.spray@redhat.com>
2014-09-29 15:04:10 +01:00
John Spray
b6ccf0d414 tasks: generalize config writing for Filesystem
Signed-off-by: John Spray <john.spray@redhat.com>
2014-09-29 15:03:17 +01:00