Commit Graph

4580 Commits

Author SHA1 Message Date
Yuri Weinstein
555d42c667 Merge pull request #602 from ceph/wip13349
White listed WAR 'failed to encode' fixes #13349
2015-10-04 14:57:50 -07:00
Yuri Weinstein
37a881aee1 White listed WAR 'failed to encode' fixes #13349
Signed-off-by: Yuri Weinstein <yweinste@redhat.com>
2015-10-03 08:27:39 -07:00
vasukulkarni
5f201625b1 Merge pull request #549 from ceph/wip-krbd-fio
Add rbd fio test to exercise block IO on all format/features
2015-10-02 11:54:26 -07:00
Vasu Kulkarni
15e890e0dc krbd fio tests using sync engine for format 2 and features 1 and 2.
Signed-off-by: Vasu Kulkarni <vasu@redhat.com>
2015-10-02 13:41:59 -04:00
Vasu Kulkarni
b5cf466ac7 krbd test using fio for format 2 and features 1 and 2.
Signed-off-by: Vasu Kulkarni <vasu@redhat.com>
2015-10-02 13:41:59 -04:00
Vasu Kulkarni
8e115c9741 rbd fio tests using rbd engine, tests on clones as well using features 1 and 13.
Signed-off-by: Vasu Kulkarni <vasu@redhat.com>
2015-10-02 13:41:59 -04:00
Vasu Kulkarni
971c5fc29c rbd io tests using fio, various fio options can be specified in yaml
test rbd or krbd using fio, can also run io on rbd clones if option is specified in yaml
various options like image-size, rbd format/features, fio io size, readwrite options can be provided in yaml.
check the docstring for exact usage.

Signed-off-by: Vasu Kulkarni <vasu@redhat.com>
2015-10-02 13:41:42 -04:00
Yuri Weinstein
138d2d7976 Merge pull request #599 from ceph/wip-hammer-x
upgrade/hammer-x/point-to-point: use different client for second rgw
2015-10-02 10:32:12 -07:00
Sage Weil
a0f0194ab2 upgrade/hammer-x/point-to-point: use different client for second rgw
Can't do 2 rgw's on the same client id.

Signed-off-by: Sage Weil <sage@redhat.com>
2015-10-02 13:24:19 -04:00
Loic Dachary
aed5a6dbc6 Merge pull request #595 from ceph/wip-quota
log-whitelist 'reached quota' for librados test.sh

Reviewed-by: Loic Dachary <ldachary@redhat.com>
2015-10-02 14:59:27 +02:00
Sage Weil
7bb1ceb3f0 log-whitelist 'reached quota' for librados test.sh
A new test verifies that we are stopped by the pool quota (and get
the right error messages or block).  See ceph.git
32962740ce.

Signed-off-by: Sage Weil <sage@redhat.com>
2015-10-02 08:33:46 -04:00
Gregory Farnum
6add6ba2df Merge pull request #461 from ceph/wip-client-oldest-tid
cephfs/test_client_limits.py: update test_client_oldest_tid

Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-10-01 22:37:15 -07:00
Sage Weil
de3c9d59df rados/singleton-nomsgr/all/full-tiering: time out rados -p
We will hit a quota and block (not error out).

See #13283
Signed-off-by: Sage Weil <sage@redhat.com>
2015-10-01 15:37:22 -04:00
Loic Dachary
487ccb964a Merge pull request #592 from ceph/wip-13000-centos-activate
ceph-deploy: always zap disk before creating an osd
2015-10-01 08:49:20 +02:00
Josh Durgin
03b4a7957e Merge pull request #580 from dachary/wip-13250-ceph-deploy-branch
ceph-deploy: fix ceph-deploy-branch config options

Reviewed-by: Josh Durgin <jdurgin@redhat.com>
2015-09-30 19:45:30 -04:00
Sage Weil
f467a98a29 tasks/ceph_manager: %d -> %s
Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-30 14:58:32 -04:00
Loic Dachary
01d48a270a ceph-deploy: always zap disk before creating an osd
The existing logic is to ceph-deploy osd create --zap-disk which will
zap the data device before preparing it. However it will not zap the
journal device (see http://tracker.ceph.com/issues/13291).

If ceph-deploy osd create fails, a fall back will zap both the data
device and the journal and try prepare again. This could work if
the device preparation and activation was synchronous and catch all
errors that could be caused by an unclean journal device. However,
the activation is asynchronous and it is entirely possible for a device
to be prepared successfully and fail to activate in the background.

The data and journal device are always zapped before calling ceph-deploy
osd create. The logic is simpler and the overhead is low.

http://tracker.ceph.com/issues/13000 Fixes: #13000

Signed-off-by: Loic Dachary <loic@dachary.org>
2015-09-30 20:57:24 +02:00
Sage Weil
5a450f8eeb rados/singleton-nomsgr/all/11429: blackhole osd.0 to close race
Blackhole filestore ops so that we ensure it doesn't complete
the pg deletions before the restart function does a clean shutdown
etc.

Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-30 14:42:05 -04:00
Sage Weil
963f5e55a9 upgrade/hammer-x/tiering: fix upgrade sequence
Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-30 09:34:26 -04:00
Sage Weil
a53a80b9f0 tasks/ceph_manager: fix logging on failed pool property
Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-30 09:24:38 -04:00
Sage Weil
415a18628c Merge pull request #591 from athanatos/wip-13234
rados: add test for 13234.yaml

Reviewed-by: Sage Weil <sage@redhat.com>
2015-09-30 08:35:10 -04:00
Samuel Just
4e9f1df514 rados: add test for 13234.yaml
Signed-off-by: Samuel Just <sjust@redhat.com>
2015-09-29 21:19:10 -07:00
John Spray
a5a361e883 Merge pull request #586 from ceph/wip-mdscap
tasks/cephfs: mds allow

Reviewed-by: John Spray <john.spray@redhat.com>
2015-09-29 10:08:20 +01:00
Gregory Farnum
79cbfc86e4 Merge pull request #575 from ceph/wip-32bits-ceph-fuse
suites/fs: test ceph-fuse with 32-bits ino_t

Reviewed-by: Greg Farnum <gfarnum@redhat.com>
2015-09-28 22:19:28 -07:00
Sage Weil
8cf3d15de7 rados/singleton-nomsgr/all/11429: grep osd.0 only
Restart can be slow enough that osd.1 and 2 finish deleting the
pgs.  Verifying one osd sees the instance is sufficient.

Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-28 13:10:45 -04:00
Sage Weil
a4f9bdbe54 tasks/cephfs: mds allow
Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-28 12:38:04 -04:00
Sage Weil
9eea6ea6f9 Merge remote-tracking branch 'gh/wip-infernalis' 2015-09-28 10:32:04 -04:00
Loic Dachary
56e95a0ebc Merge pull request #583 from dachary/wip-buildpackages-flock
buildpackages: flock must --close
2015-09-28 02:03:54 +02:00
Loic Dachary
0b130d7f53 Merge pull request #584 from dachary/wip-buildpackages-centos-6
buildpackages: user-data must be per os-type/os-version
2015-09-28 02:01:30 +02:00
Loic Dachary
05228e1e5c buildpackages: user-data must be per os-type/os-version
CentOS 6.5 needs to install a package and reboot to grow the root file
system. Instead of assuming a common user-data.txt file can fit all
Operating Systems, make one user data per os-type/os-version combination.

Signed-off-by: Loic Dachary <loic@dachary.org>
2015-09-27 23:42:58 +02:00
Kefu Chai
d58a0b00e8 add cache tiering test for hammer-x
Signed-off-by: Kefu Chai <tchaikov@gmail.com>
2015-09-27 08:24:38 -04:00
Sage Weil
8941a8b871 rados/singleton-nomsgr/all/11439: fix greps
Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-26 21:11:31 -04:00
Loic Dachary
204e27fecd buildpackages: flock must --close
The process run by flock must not inherit the file descriptor because
this will cause the lock to be held forever should the command survive
the call to flock. This is for instance the case for the ssh-agent.

Signed-off-by: Loic Dachary <loic@dachary.org>
2015-09-26 15:23:03 +02:00
Loic Dachary
8bc89b6517 Merge pull request #581 from dachary/wip-11883-git-url
buildpackages: get Ceph git url via get_ceph_git_url()
2015-09-26 13:44:56 +02:00
Loic Dachary
78b2a9aa74 buildpackages: get Ceph git url via get_ceph_git_url()
Instead of relying on git_base_url, use the get_ceph_git_url() to obtain
the URL of the Ceph repository to use with git clone. This allows the
user to override it via the git_ceph_url configuration option and the
--git-ceph-url command line option to teuthology-openstack.

http://tracker.ceph.com/issues/11883 Refs: #11883

Signed-off-by: Loic Dachary <loic@dachary.org>
2015-09-26 13:11:36 +02:00
Loic Dachary
fdcfb79b3c ceph-deploy: fix ceph-deploy-branch config options
The config paramter of download_ceph_deploy does not have a ceph-deploy
item, therefore the ceph-deploy-branch parameter is always assumed to be
master.

Signed-off-by: Loic Dachary <loic@dachary.org>
2015-09-26 02:16:46 +02:00
Sage Weil
9e95c7430a upgrade/hammer-x/parallel: set infernalis settings after upgrade
Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-24 23:37:35 -04:00
Sage Weil
910770c318 releases/infernalis.yaml: final steps after infernalis upgrade
Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-24 23:37:35 -04:00
Sage Weil
ecc504fea0 tasks/divergent_priors2: wait for osd to start before tell
Otherwise we can get

2015-09-24T19:22:15.191 INFO:teuthology.orchestra.run.mira080.stderr:Error ENXIO: problem getting command descriptions from osd.1

Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-24 23:34:17 -04:00
Sage Weil
ea5cea8e60 rados/singleton-nomsgr/all/11429: more reliably create orphan pgs
Make sure osds have the latest map with a pg ls

Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-24 23:11:41 -04:00
Sage Weil
e01dedb65e Merge pull request #574 from ceph/wip-13098
rados/singleton-nomsgr/all/full-tiering: test full cache tier
2015-09-24 22:56:53 -04:00
Sage Weil
cff9bddba4 rados/singleton-nomsgr/all/full-tiering: test full cache tier
Verify fix for #13098

Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-24 22:55:56 -04:00
Zack Cerza
3736ca2a39 Merge pull request #577 from ceph/wip-get-prop
tolerate pg_num get failure
2015-09-24 10:20:44 -06:00
Sage Weil
0e2814d81e tasks/ceph_manager: ignore failure getting pg_num
Otherwise, we may fail while racing with a workload that deletes a pool:

2015-09-23T15:01:52.855 INFO:tasks.workunit.client.1.vpm128.stdout:[ RUN      ] LibRadosTwoPoolsPP.PromoteSnapTrimRace
2015-09-23T15:01:53.892 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get .rgw pg_num'
2015-09-23T15:01:54.206 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get .rgw.gc pg_num'
2015-09-23T15:01:54.462 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get .users.uid pg_num'
2015-09-23T15:01:54.696 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get .users.email pg_num'
2015-09-23T15:01:55.006 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get .users pg_num'
2015-09-23T15:01:55.296 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get .rgw.buckets.index pg_num'
2015-09-23T15:01:55.523 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get .log pg_num'
2015-09-23T15:01:55.752 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get .usage pg_num'
2015-09-23T15:01:56.188 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get .rgw.buckets.extra pg_num'
2015-09-23T15:01:56.625 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get test-rados-api-vpm128-17360-6 pg_num'
2015-09-23T15:01:56.928 INFO:teuthology.orchestra.run.vpm176:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool get test-rados-api-vpm128-17360-13 pg_num'
2015-09-23T15:01:57.193 INFO:teuthology.orchestra.run.vpm176.stderr:Error ENOENT: unrecognized pool 'test-rados-api-vpm128-17360-13'
2015-09-23T15:01:57.206 ERROR:teuthology.parallel:Exception in parallel execution
Traceback (most recent call last):
...

Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-24 12:19:07 -04:00
Loic Dachary
ce0033544f Merge pull request #576 from ceph/wip-13031-packages
buildpackages: make rpm packages


Reviewed-by: Joe Julian <jjulian@io.com>
2015-09-23 21:37:25 +02:00
Loic Dachary
7e3a271a89 buildpackages: make rpm packages
Signed-off-by: Loic Dachary <loic@dachary.org>
2015-09-23 21:11:26 +02:00
Sage Weil
97d64b4e86 Merge pull request #573 from ceph/wip-hammer-x
hammer-x fixes

Reviewed-by: Loic Dachary <ldachary@redhat.com>
2015-09-23 15:06:55 -04:00
Sage Weil
8087095fa6 upgrade/hammer-x/stress-split-erasure-code-x86_64: specify arch
Signed-off-by: Sage Weil <sage@redhat.com>
2015-09-23 15:06:19 -04:00
Yan, Zheng
7fbd70b697 suites/fs: test ceph-fuse with 32-bits ino_t
Signed-off-by: Yan, Zheng <zyan@redhat.com>
2015-09-23 16:01:07 +08:00
Sage Weil
a653f6c02e Merge pull request #571 from dachary/wip-isa-arch
erasure-code: isolate isa plugin arch constraint
2015-09-22 13:40:50 -04:00