diff --git a/suites/rados/upgrade/hammer-x-singleton/% b/suites/rados/upgrade/hammer-x-singleton/% new file mode 100644 index 00000000000..e69de29bb2d diff --git a/suites/rados/upgrade/hammer-x-singleton/0-cluster/start.yaml b/suites/rados/upgrade/hammer-x-singleton/0-cluster/start.yaml new file mode 100644 index 00000000000..a7dbead143f --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/0-cluster/start.yaml @@ -0,0 +1,17 @@ +overrides: + ceph: + conf: + mon: + mon warn on legacy crush tunables: false +roles: +- - mon.a + - mon.b + - mon.c + - mds.a + - osd.0 + - osd.1 + - osd.2 +- - osd.3 + - osd.4 + - osd.5 +- - client.0 diff --git a/suites/rados/upgrade/hammer-x-singleton/1-hammer-install/hammer.yaml b/suites/rados/upgrade/hammer-x-singleton/1-hammer-install/hammer.yaml new file mode 100644 index 00000000000..6c2470dffe1 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/1-hammer-install/hammer.yaml @@ -0,0 +1,7 @@ +tasks: +- install: + branch: hammer +- print: "**** done install hammer" +- ceph: + fs: xfs +- print: "**** done ceph" diff --git a/suites/rados/upgrade/hammer-x-singleton/2-partial-upgrade/firsthalf.yaml b/suites/rados/upgrade/hammer-x-singleton/2-partial-upgrade/firsthalf.yaml new file mode 100644 index 00000000000..2490f76d638 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/2-partial-upgrade/firsthalf.yaml @@ -0,0 +1,7 @@ +tasks: +- install.upgrade: + osd.0: +- print: "**** done install.upgrade osd.0" +- ceph.restart: + daemons: [osd.0, osd.1, osd.2, osd.3] +- print: "**** done ceph.restart 1st half" diff --git a/suites/rados/upgrade/hammer-x-singleton/3-thrash/default.yaml b/suites/rados/upgrade/hammer-x-singleton/3-thrash/default.yaml new file mode 100644 index 00000000000..d99d0c87420 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/3-thrash/default.yaml @@ -0,0 +1,12 @@ +overrides: + ceph: + log-whitelist: + - wrongly marked me down + - objects unfound and apparently lost + - log bound mismatch +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgpnum_fix: 1 +- print: "**** done thrashosds 3-thrash" diff --git a/suites/rados/upgrade/hammer-x-singleton/4-mon/mona.yaml b/suites/rados/upgrade/hammer-x-singleton/4-mon/mona.yaml new file mode 100644 index 00000000000..7c75c102fc4 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/4-mon/mona.yaml @@ -0,0 +1,6 @@ +tasks: +- ceph.restart: + daemons: [mon.a] + wait-for-healthy: false + wait-for-osds-up: true +- print: "**** done ceph.restart mon.a" diff --git a/suites/rados/upgrade/hammer-x-singleton/5-workload/+ b/suites/rados/upgrade/hammer-x-singleton/5-workload/+ new file mode 100644 index 00000000000..e69de29bb2d diff --git a/suites/rados/upgrade/hammer-x-singleton/5-workload/rbd-cls.yaml b/suites/rados/upgrade/hammer-x-singleton/5-workload/rbd-cls.yaml new file mode 100644 index 00000000000..1bcd1b8ccad --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/5-workload/rbd-cls.yaml @@ -0,0 +1,7 @@ +tasks: +- workunit: + branch: hammer + clients: + client.0: + - cls/test_cls_rbd.sh +- print: "**** done cls/test_cls_rbd.sh 5-workload" diff --git a/suites/rados/upgrade/hammer-x-singleton/5-workload/rbd-import-export.yaml b/suites/rados/upgrade/hammer-x-singleton/5-workload/rbd-import-export.yaml new file mode 100644 index 00000000000..882d979c89a --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/5-workload/rbd-import-export.yaml @@ -0,0 +1,9 @@ +tasks: +- workunit: + branch: hammer + clients: + client.0: + - rbd/import_export.sh + env: + RBD_CREATE_ARGS: --new-format +- print: "**** done rbd/import_export.sh 5-workload" diff --git a/suites/rados/upgrade/hammer-x-singleton/5-workload/readwrite.yaml b/suites/rados/upgrade/hammer-x-singleton/5-workload/readwrite.yaml new file mode 100644 index 00000000000..922e61fdfc0 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/5-workload/readwrite.yaml @@ -0,0 +1,11 @@ +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 500 + write_append_excl: false + op_weights: + read: 45 + write: 45 + delete: 10 +- print: "**** done rados/readwrite 5-workload" diff --git a/suites/rados/upgrade/hammer-x-singleton/5-workload/snaps-few-objects.yaml b/suites/rados/upgrade/hammer-x-singleton/5-workload/snaps-few-objects.yaml new file mode 100644 index 00000000000..33c820d1277 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/5-workload/snaps-few-objects.yaml @@ -0,0 +1,14 @@ +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + write_append_excl: false + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 +- print: "**** done rados/snaps-few-objects 5-workload" diff --git a/suites/rados/upgrade/hammer-x-singleton/6-next-mon/monb.yaml b/suites/rados/upgrade/hammer-x-singleton/6-next-mon/monb.yaml new file mode 100644 index 00000000000..22e87c7ad21 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/6-next-mon/monb.yaml @@ -0,0 +1,6 @@ +tasks: +- ceph.restart: + daemons: [mon.b] + wait-for-healthy: false + wait-for-osds-up: true +- print: "**** done ceph.restart mon.b 6-next-mon" diff --git a/suites/rados/upgrade/hammer-x-singleton/7-workload/+ b/suites/rados/upgrade/hammer-x-singleton/7-workload/+ new file mode 100644 index 00000000000..e69de29bb2d diff --git a/suites/rados/upgrade/hammer-x-singleton/7-workload/radosbench.yaml b/suites/rados/upgrade/hammer-x-singleton/7-workload/radosbench.yaml new file mode 100644 index 00000000000..3d87bb1c2c5 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/7-workload/radosbench.yaml @@ -0,0 +1,5 @@ +tasks: +- radosbench: + clients: [client.0] + time: 1800 +- print: "**** done radosbench 7-workload" diff --git a/suites/rados/upgrade/hammer-x-singleton/7-workload/rbd_api.yaml b/suites/rados/upgrade/hammer-x-singleton/7-workload/rbd_api.yaml new file mode 100644 index 00000000000..19a24c2a7f5 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/7-workload/rbd_api.yaml @@ -0,0 +1,7 @@ +tasks: +- workunit: + branch: hammer + clients: + client.0: + - rbd/test_librbd.sh +- print: "**** done rbd/test_librbd.sh 7-workload" diff --git a/suites/rados/upgrade/hammer-x-singleton/8-next-mon/monc.yaml b/suites/rados/upgrade/hammer-x-singleton/8-next-mon/monc.yaml new file mode 100644 index 00000000000..61253685890 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/8-next-mon/monc.yaml @@ -0,0 +1,8 @@ +tasks: +- ceph.restart: + daemons: [mon.c] + wait-for-healthy: false + wait-for-osds-up: true +- print: "**** done ceph.restart mon.c 8-next-mon" +- ceph.wait_for_mon_quorum: [a, b, c] +- print: "**** done wait_for_mon_quorum 8-next-mon" diff --git a/suites/rados/upgrade/hammer-x-singleton/9-workload/+ b/suites/rados/upgrade/hammer-x-singleton/9-workload/+ new file mode 100644 index 00000000000..e69de29bb2d diff --git a/suites/rados/upgrade/hammer-x-singleton/9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml b/suites/rados/upgrade/hammer-x-singleton/9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml new file mode 100644 index 00000000000..7714312c339 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml @@ -0,0 +1,31 @@ +# +# k=3 implies a stripe_width of 1376*3 = 4128 which is different from +# the default value of 4096 It is also not a multiple of 1024*1024 and +# creates situations where rounding rules during recovery becomes +# necessary. +# +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + ec_pool: true + write_append_excl: false + erasure_code_profile: + name: jerasure31profile + plugin: jerasure + k: 3 + m: 1 + technique: reed_sol_van + ruleset-failure-domain: osd + op_weights: + read: 100 + write: 0 + append: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 + setattr: 25 + rmattr: 25 diff --git a/suites/rados/upgrade/hammer-x-singleton/9-workload/rbd-python.yaml b/suites/rados/upgrade/hammer-x-singleton/9-workload/rbd-python.yaml new file mode 100644 index 00000000000..0ea5639e6bb --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/9-workload/rbd-python.yaml @@ -0,0 +1,7 @@ +tasks: +- workunit: + branch: hammer + clients: + client.0: + - rbd/test_librbd_python.sh +- print: "**** done rbd/test_librbd_python.sh 9-workload" diff --git a/suites/rados/upgrade/hammer-x-singleton/9-workload/rgw-swift.yaml b/suites/rados/upgrade/hammer-x-singleton/9-workload/rgw-swift.yaml new file mode 100644 index 00000000000..8f1416082bb --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/9-workload/rgw-swift.yaml @@ -0,0 +1,9 @@ +tasks: +- rgw: + client.0: + default_idle_timeout: 300 +- print: "**** done rgw 9-workload" +- swift: + client.0: + rgw_server: client.0 +- print: "**** done swift 9-workload" diff --git a/suites/rados/upgrade/hammer-x-singleton/9-workload/snaps-many-objects.yaml b/suites/rados/upgrade/hammer-x-singleton/9-workload/snaps-many-objects.yaml new file mode 100644 index 00000000000..39b44f665f1 --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/9-workload/snaps-many-objects.yaml @@ -0,0 +1,13 @@ +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 500 + write_append_excl: false + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 diff --git a/suites/rados/upgrade/hammer-x-singleton/9-workload/test_cache-pool-snaps.yaml b/suites/rados/upgrade/hammer-x-singleton/9-workload/test_cache-pool-snaps.yaml new file mode 100644 index 00000000000..528050d4b4a --- /dev/null +++ b/suites/rados/upgrade/hammer-x-singleton/9-workload/test_cache-pool-snaps.yaml @@ -0,0 +1,36 @@ +overrides: + ceph: + log-whitelist: + - must scrub before tier agent can activate +tasks: +workunit: + sequential: + - exec: + client.0: + - ceph osd pool create base 4 + - ceph osd pool create cache 4 + - ceph osd tier add base cache + - ceph osd tier cache-mode cache writeback + - ceph osd tier set-overlay base cache + - ceph osd pool set cache hit_set_type bloom + - ceph osd pool set cache hit_set_count 8 + - ceph osd pool set cache hit_set_period 3600 + - ceph osd pool set cache target_max_objects 250 + - rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + pool_snaps: true + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 + flush: 50 + try_flush: 50 + evict: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + - print: "**** done test_cache-pool-snaps 9-workload"