copyed upgrade/infernalis suite to master branch

Signed-off-by: Yuri Weinstein <yweinste@redhat.com>
This commit is contained in:
Yuri Weinstein 2016-05-04 08:46:29 -07:00
parent 1fde198f97
commit 7aa2ec1f9e
21 changed files with 500 additions and 0 deletions

View File

View File

@ -0,0 +1,32 @@
meta:
- desc: |
Run ceph on two nodes, using one of them as a client,
with a separate client-only node.
Use xfs beneath the osds.
overrides:
ceph:
log-whitelist:
- scrub
- scrub mismatch
- ScrubResult
- wrongly marked me down
- objects unfound and apparently lost
- log bound mismatch
- failed to encode map
- soft lockup
- detected stalls on CPUs
fs: xfs
roles:
- - mon.a
- mds.a
- osd.0
- osd.1
- osd.2
- - mon.b
- mon.c
- osd.3
- osd.4
- osd.5
- client.0
- - client.1
- client.2

View File

@ -0,0 +1,13 @@
meta:
- desc: |
install ceph/hammer latest release
run workload and upgrade-sequence in parallel
tasks:
- install:
branch: hammer
- print: "**** done latest hammer install"
- ceph:
- parallel:
- workload
- upgrade-sequence
- print: "**** done parallel hammer"

View File

@ -0,0 +1,13 @@
meta:
- desc: |
install ceph/infernalis v9.2.0
run workload and upgrade-sequence in parallel
tasks:
- install:
tag: v9.2.0
- print: "**** done v9.2.0 install"
- ceph:
- parallel:
- workload
- upgrade-sequence
- print: "**** done parallel v9.2.0"

View File

@ -0,0 +1,13 @@
meta:
- desc: |
install ceph/infernalis v9.2.1
run workload and upgrade-sequence in parallel
tasks:
- install:
tag: v9.2.1
- print: "**** done v9.2.1 install"
- ceph:
- parallel:
- workload
- upgrade-sequence
- print: "**** done parallel v9.2.1"

View File

@ -0,0 +1,13 @@
meta:
- desc: |
run a cephfs stress test
mount ceph-fuse on client.0 before running workunit
workload:
sequential:
- ceph-fuse:
- print: "**** done ceph-fuse 2-workload"
- workunit:
clients:
client.0:
- suites/blogbench.sh
- print: "**** done suites/blogbench.sh 2-workload"

View File

@ -0,0 +1,12 @@
meta:
- desc: |
run basic import/export cli tests for rbd
workload:
sequential:
- workunit:
clients:
client.1:
- rbd/import_export.sh
env:
RBD_CREATE_ARGS: --new-format
- print: "**** done rbd/import_export.sh 2-workload"

View File

@ -0,0 +1,11 @@
meta:
- desc: |
add test details here
run s3tests
workload:
sequential:
- rgw: [client.1]
- s3tests:
client.0:
force-branch: hammer
rgw_server: client.1

View File

@ -0,0 +1,16 @@
meta:
- desc: |
run randomized correctness test for rados operations
workload:
sequential:
- rados:
clients: [client.0]
ops: 2000
objects: 50
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50

View File

@ -0,0 +1,41 @@
meta:
- desc: |
upgrade the ceph cluster,
ordering: mon, mds, osd
upgrade-sequence:
sequential:
- install.upgrade:
mon.a:
mon.b:
- print: "**** done install.upgrade infernalis"
- ceph.restart: [mon.a]
- sleep:
duration: 60
- ceph.restart: [mon.b]
- sleep:
duration: 60
- ceph.restart: [mon.c]
- sleep:
duration: 60
- ceph.restart: [mds.a]
- sleep:
duration: 60
- ceph.restart: [osd.0]
- sleep:
duration: 30
- ceph.restart: [osd.1]
- sleep:
duration: 30
- ceph.restart: [osd.2]
- sleep:
duration: 30
- ceph.restart: [osd.3]
- sleep:
duration: 30
- ceph.restart: [osd.4]
- sleep:
duration: 30
- ceph.restart: [osd.5]
- sleep:
duration: 30
- print: "**** done ceph.restart all"

View File

@ -0,0 +1,41 @@
meta:
- desc: |
upgrade the ceph cluster,
ordering: osd, mon, mds
upgrade-sequence:
sequential:
- install.upgrade:
mon.a:
mon.b:
- print: "**** done install.upgrade infernalis"
- ceph.restart: [osd.0]
- sleep:
duration: 30
- ceph.restart: [osd.1]
- sleep:
duration: 30
- ceph.restart: [osd.2]
- sleep:
duration: 30
- ceph.restart: [osd.3]
- sleep:
duration: 30
- ceph.restart: [osd.4]
- sleep:
duration: 30
- ceph.restart: [osd.5]
- sleep:
duration: 60
- ceph.restart: [mon.a]
- sleep:
duration: 60
- ceph.restart: [mon.b]
- sleep:
duration: 60
- ceph.restart: [mon.c]
- sleep:
duration: 60
- ceph.restart: [mds.a]
- sleep:
duration: 60
- print: "**** done ceph.restart all"

View File

@ -0,0 +1,16 @@
meta:
- desc: |
randomly kill and revive monitors
tasks:
- mon_thrash:
revive_delay: 20
thrash_delay: 1
- print: "**** done mon_thrash 4-workload"
- ceph-fuse:
- print: "**** done ceph-fuse 4-workload"
- workunit:
clients:
client.0:
- suites/dbench.sh
- print: "**** done suites/dbench.sh 4-workload"

View File

@ -0,0 +1,18 @@
meta:
- desc: |
randomly kill and revive osds
increasing the number of pgs at random times
run cephfs stress test
tasks:
- sequential:
- thrashosds:
timeout: 1200
chance_pgnum_grow: 1
chance_pgpnum_fix: 1
- print: "**** done thrashosds 4-workload"
- workunit:
clients:
client.0:
- suites/iogen.sh
- print: "**** done suites/iogen.sh 4-workload"

View File

@ -0,0 +1,16 @@
meta:
- desc: |
run randomized correctness test for rados operations
tasks:
- sequential:
- rados:
clients: [client.0]
ops: 2000
objects: 50
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50

View File

@ -0,0 +1,9 @@
tasks:
- sequential:
- rgw: [client.2]
- print: "**** done rgw: [client.2] 4-workload"
- s3tests:
client.2:
force-branch: hammer
rgw_server: client.2
- print: "**** done s3tests 4-workload"

View File

@ -0,0 +1 @@
../../../../distros/supported/

View File

@ -0,0 +1 @@
../../../../distros/supported

View File

@ -0,0 +1,234 @@
overrides:
ceph:
log-whitelist:
- reached quota
- scrub
- osd_map_max_advance
fs: xfs
conf:
mon:
mon debug unsafe allow tier with nonempty snaps: true
osd:
osd map max advance: 1000
roles:
- - mon.a
- mds.a
- osd.0
- osd.1
- osd.2
- - mon.b
- mon.c
- osd.3
- osd.4
- osd.5
- client.0
- - client.1
tasks:
- print: "**** v9.2.0 about to install"
- install:
tag: v9.2.0
- print: "**** done v9.2.0 install"
- ceph:
fs: xfs
- print: "**** done ceph xfs"
- sequential:
- workload
- print: "**** done workload v9.2.0"
- parallel:
- workload1
- upgrade-sequence1
- print: "**** done parallel v9.2.1"
###- parallel:
### - workloadX
### - upgrade-sequenceX
###- print: "**** done parallel v9.2.X"
- parallel:
- workload_infernalis
- upgrade-sequence_infernalis
- print: "**** done parallel infernalis branch"
#######################
workload:
sequential:
- workunit:
clients:
client.0:
- suites/blogbench.sh
- print: "**** done suites/blogbench.sh workload"
workload1:
sequential:
- workunit:
clients:
client.0:
- rados/load-gen-big.sh
- print: "**** done rados/load-gen-big.sh workload1"
- workunit:
clients:
client.0:
- rados/test.sh
- cls
- print: "**** done rados/test.sh & cls workload1"
- workunit:
clients:
client.0:
- rbd/test_librbd.sh
- print: "**** done rbd/test_librbd.sh workload1"
upgrade-sequence1:
sequential:
- install.upgrade:
mon.a:
tag: v9.2.1
mon.b:
tag: v9.2.1
client.1:
tag: v9.2.1
- print: "**** done v9.2.1 install.upgrade"
- ceph.restart: [mon.a]
- sleep:
duration: 60
- ceph.restart: [mon.b]
- sleep:
duration: 60
- ceph.restart: [mon.c]
- sleep:
duration: 60
- ceph.restart: [mds.a]
- sleep:
duration: 60
- ceph.restart: [osd.0]
- sleep:
duration: 30
- ceph.restart: [osd.1]
- sleep:
duration: 30
- ceph.restart: [osd.2]
- sleep:
duration: 30
- ceph.restart: [osd.3]
- sleep:
duration: 30
- ceph.restart: [osd.4]
- sleep:
duration: 30
- ceph.restart: [osd.5]
- sleep:
duration: 30
- print: "**** done ceph.restart all mon/mds/osd v0.94.1"
### =====
###workloadX:
### sequential:
### - workunit:
### clients:
### client.0:
### - rados/load-gen-big.sh
### - print: "**** done rados/load-gen-big.sh workload1"
### - workunit:
### clients:
### client.0:
### - rados/test.sh
### - cls
### - print: "**** done rados/test.sh & cls workload1"
### - workunit:
### clients:
### client.0:
### - rbd/test_librbd.sh
### - print: "**** done rbd/test_librbd.sh workload1"
###upgrade-sequenceX:
### sequential:
### - install.upgrade:
### mon.a:
### tag: v9.2.X
### mon.b:
### tag: v9.2.X
### client.1:
### tag: v9.2.X
### - print: "**** done v9.2.X install.upgrade"
### - ceph.restart: [mon.a]
### - sleep:
### duration: 60
### - ceph.restart: [mon.b]
### - sleep:
### duration: 60
### - ceph.restart: [mon.c]
### - sleep:
### duration: 60
### - ceph.restart: [mds.a]
### - sleep:
### duration: 60
### - ceph.restart: [osd.0]
### - sleep:
### duration: 30
### - ceph.restart: [osd.1]
### - sleep:
### duration: 30
### - ceph.restart: [osd.2]
### - sleep:
### duration: 30
### - ceph.restart: [osd.3]
### - sleep:
### duration: 30
### - ceph.restart: [osd.4]
### - sleep:
### duration: 30
### - ceph.restart: [osd.5]
### - sleep:
### duration: 30
### - print: "**** done ceph.restart all mon/mds/osd v0.94.X"
workload_infernalis:
sequential:
- rgw: [client.0]
- print: "**** done rgw workload_infernalis"
- s3tests:
client.0:
# for s3tests ceph-infernalis should be used
# for new similar tests should check with Yehuda about the branch name
force-branch: ceph-infernalis
rgw_server: client.0
- print: "**** done s3tests workload_infernalis"
upgrade-sequence_infernalis:
sequential:
- install.upgrade:
mon.a:
branch: infernalis
mon.b:
branch: infernalis
client.1:
branch: infernalis
- print: "**** done branch: infernalis install.upgrade"
- ceph.restart: [mds.a]
- sleep:
duration: 60
- ceph.restart: [osd.0]
- sleep:
duration: 30
- ceph.restart: [osd.1]
- sleep:
duration: 30
- ceph.restart: [osd.2]
- sleep:
duration: 30
- ceph.restart: [osd.3]
- sleep:
duration: 30
- ceph.restart: [osd.4]
- sleep:
duration: 30
- ceph.restart: [osd.5]
- sleep:
duration: 60
- ceph.restart: [mon.a]
- sleep:
duration: 60
- ceph.restart: [mon.b]
- sleep:
duration: 60
- ceph.restart: [mon.c]
- sleep:
duration: 60
- print: "**** done ceph.restart all infernalis current branch mds/osd/mon"