From: Yuri Weinstein Date: Fri, 18 Jan 2019 20:53:05 +0000 (-0800) Subject: qa/tests: changed to `supported' distro and moved all `p2p` suites under the same... X-Git-Tag: v12.2.11~4^2 X-Git-Url: http://git-server-git.apps.pok.os.sepia.ceph.com/?a=commitdiff_plain;h=refs%2Fpull%2F26037%2Fhead;p=ceph.git qa/tests: changed to `supported' distro and moved all `p2p` suites under the same folder Signed-off-by: Yuri Weinstein --- diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/% b/qa/suites/upgrade/luminous-p2p-stress-split/% deleted file mode 100644 index e69de29bb2d1..000000000000 diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/+ b/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/+ deleted file mode 100644 index e69de29bb2d1..000000000000 diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/.qa b/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/.qa deleted file mode 120000 index a602a0353e75..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/.qa +++ /dev/null @@ -1 +0,0 @@ -../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/openstack.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/openstack.yaml deleted file mode 100644 index a0d5c20195be..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/openstack.yaml +++ /dev/null @@ -1,6 +0,0 @@ -openstack: - - machine: - disk: 100 # GB - - volumes: # attached to each instance - count: 3 - size: 30 # GB diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/start.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/start.yaml deleted file mode 100644 index 4f40219b5267..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/0-cluster/start.yaml +++ /dev/null @@ -1,20 +0,0 @@ -meta: -- desc: | - Run ceph on two nodes, - with a separate client-only node. - Use xfs beneath the osds. -overrides: - ceph: - fs: xfs -roles: -- - mon.a - - mon.b - - mon.c - - mgr.x - - osd.0 - - osd.1 - - osd.2 -- - osd.3 - - osd.4 - - osd.5 -- - client.0 diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/1-ceph-install/luminous.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/1-ceph-install/luminous.yaml deleted file mode 100644 index b66e0ca99f06..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/1-ceph-install/luminous.yaml +++ /dev/null @@ -1,19 +0,0 @@ -meta: -- desc: install ceph/luminous latest -tasks: -- install: - tag: v12.2.10 - exclude_packages: ['librados3'] - extra_packages: ['librados2'] -- print: "**** done install luminous v12.2.10" -- ceph: -- exec: - osd.0: - - ceph osd require-osd-release luminous - - ceph osd set-require-min-compat-client luminous -- print: "**** done ceph" -overrides: - ceph: - conf: - mon: - mon warn on osd down out interval zero: false diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/1.1-pg-log-overrides/normal_pg_log.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/1.1-pg-log-overrides/normal_pg_log.yaml deleted file mode 100644 index 8b137891791f..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/1.1-pg-log-overrides/normal_pg_log.yaml +++ /dev/null @@ -1 +0,0 @@ - diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/1.1-pg-log-overrides/short_pg_log.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/1.1-pg-log-overrides/short_pg_log.yaml deleted file mode 100644 index 20cc101dea1b..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/1.1-pg-log-overrides/short_pg_log.yaml +++ /dev/null @@ -1,6 +0,0 @@ -overrides: - ceph: - conf: - global: - osd_min_pg_log_entries: 1 - osd_max_pg_log_entries: 2 diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/2-partial-upgrade/.qa b/qa/suites/upgrade/luminous-p2p-stress-split/2-partial-upgrade/.qa deleted file mode 120000 index a602a0353e75..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/2-partial-upgrade/.qa +++ /dev/null @@ -1 +0,0 @@ -../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/2-partial-upgrade/firsthalf.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/2-partial-upgrade/firsthalf.yaml deleted file mode 100644 index a73b87beb4c5..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/2-partial-upgrade/firsthalf.yaml +++ /dev/null @@ -1,17 +0,0 @@ -meta: -- desc: | - install upgrade ceph/-x on one node only - 1st half - restart : osd.0,1,2 -tasks: -- install.upgrade: - osd.0: -- print: "**** done install.upgrade osd.0" -- ceph.restart: - daemons: [mon.a,mon.b,mon.c,osd.0, osd.1, osd.2] -- print: "**** done ceph.restart 1st half" -- exec: - osd.0: - - ceph osd set pglog_hardlimit && exit 1 || true - - ceph osd dump --format=json-pretty | grep "flags" -- print: "**** try to set pglog_hardlimit, should not succeed" diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/3-thrash/.qa b/qa/suites/upgrade/luminous-p2p-stress-split/3-thrash/.qa deleted file mode 120000 index a602a0353e75..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/3-thrash/.qa +++ /dev/null @@ -1 +0,0 @@ -../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/3-thrash/default.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/3-thrash/default.yaml deleted file mode 100644 index b3fddefc7a35..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/3-thrash/default.yaml +++ /dev/null @@ -1,25 +0,0 @@ -meta: -- desc: | - randomly kill and revive osd - small chance to increase the number of pgs -overrides: - ceph: - log-whitelist: - - but it is still running - - wrongly marked me down - - objects unfound and apparently lost - - log bound mismatch -tasks: -- parallel: - - stress-tasks -stress-tasks: -- thrashosds: - timeout: 1200 - chance_pgnum_grow: 1 - chance_pgpnum_fix: 1 - chance_thrash_cluster_full: 0 - chance_thrash_pg_upmap: 0 - chance_thrash_pg_upmap_items: 0 - disable_objectstore_tool_tests: true - chance_force_recovery: 0 -- print: "**** done thrashosds 3-thrash" diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/+ b/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/+ deleted file mode 100644 index e69de29bb2d1..000000000000 diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/.qa b/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/.qa deleted file mode 120000 index a602a0353e75..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/.qa +++ /dev/null @@ -1 +0,0 @@ -../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/radosbench.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/radosbench.yaml deleted file mode 100644 index 626ae8ea6653..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/radosbench.yaml +++ /dev/null @@ -1,40 +0,0 @@ -meta: -- desc: | - run randomized correctness test for rados operations - generate write load with rados bench -stress-tasks: -- full_sequential: - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 -- print: "**** done radosbench 7-workload" diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/rbd-cls.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/rbd-cls.yaml deleted file mode 100644 index f8cc4d8ac32b..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/rbd-cls.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - run basic cls tests for rbd -stress-tasks: -- workunit: - branch: luminous - clients: - client.0: - - cls/test_cls_rbd.sh -- print: "**** done cls/test_cls_rbd.sh 5-workload" diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/rbd-import-export.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/rbd-import-export.yaml deleted file mode 100644 index 30a677af65cf..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/rbd-import-export.yaml +++ /dev/null @@ -1,12 +0,0 @@ -meta: -- desc: | - run basic import/export cli tests for rbd -stress-tasks: -- workunit: - branch: luminous - clients: - client.0: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --new-format -- print: "**** done rbd/import_export.sh 5-workload" diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/rbd_api.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/rbd_api.yaml deleted file mode 100644 index 9079aa33b897..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/rbd_api.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - librbd C and C++ api tests -stress-tasks: -- workunit: - branch: luminous - clients: - client.0: - - rbd/test_librbd.sh -- print: "**** done rbd/test_librbd.sh 7-workload" diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/readwrite.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/readwrite.yaml deleted file mode 100644 index 41e34d6d7d32..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/readwrite.yaml +++ /dev/null @@ -1,16 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool, - using only reads, writes, and deletes -stress-tasks: -- full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 500 - write_append_excl: false - op_weights: - read: 45 - write: 45 - delete: 10 -- print: "**** done rados/readwrite 5-workload" diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/snaps-few-objects.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/snaps-few-objects.yaml deleted file mode 100644 index f56d0de0f745..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/4-workload/snaps-few-objects.yaml +++ /dev/null @@ -1,18 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshot operations -stress-tasks: -- full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 -- print: "**** done rados/snaps-few-objects 5-workload" diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/5-finish-upgrade.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/5-finish-upgrade.yaml deleted file mode 100644 index 9d5a96c6cecd..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/5-finish-upgrade.yaml +++ /dev/null @@ -1,14 +0,0 @@ -tasks: -- install.upgrade: - osd.3: - client.0: -- ceph.restart: - daemons: [osd.3, osd.4, osd.5] - wait-for-healthy: false - wait-for-osds-up: true -- exec: - osd.0: - - ceph osd set pglog_hardlimit - - ceph osd dump --format=json-pretty | grep "flags" -- print: "**** try to set pglog_hardlimit again, should succeed" - diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/+ b/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/+ deleted file mode 100644 index e69de29bb2d1..000000000000 diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/.qa b/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/.qa deleted file mode 120000 index a602a0353e75..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/.qa +++ /dev/null @@ -1 +0,0 @@ -../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/rbd-python.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/rbd-python.yaml deleted file mode 100644 index 56ba21d7a180..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/rbd-python.yaml +++ /dev/null @@ -1,9 +0,0 @@ -meta: -- desc: | - librbd python api tests -tasks: -- workunit: - clients: - client.0: - - rbd/test_librbd_python.sh -- print: "**** done rbd/test_librbd_python.sh 9-workload" diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/rgw-swift.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/rgw-swift.yaml deleted file mode 100644 index 76e5d6fc2315..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/rgw-swift.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - swift api tests for rgw -tasks: -- rgw: - client.0: -- print: "**** done rgw 9-workload" -- swift: - client.0: - rgw_server: client.0 -- print: "**** done swift 9-workload" diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/snaps-many-objects.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/snaps-many-objects.yaml deleted file mode 100644 index 805bf97c3792..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/7-final-workload/snaps-many-objects.yaml +++ /dev/null @@ -1,16 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshot operations -tasks: -- rados: - clients: [client.0] - ops: 4000 - objects: 500 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/distros b/qa/suites/upgrade/luminous-p2p-stress-split/distros deleted file mode 120000 index dd0d7f1d5bdc..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/distros +++ /dev/null @@ -1 +0,0 @@ -../../../distros/supported/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p-stress-split/thrashosds-health.yaml b/qa/suites/upgrade/luminous-p2p-stress-split/thrashosds-health.yaml deleted file mode 120000 index ebf7f34f39bc..000000000000 --- a/qa/suites/upgrade/luminous-p2p-stress-split/thrashosds-health.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../tasks/thrashosds-health.yaml \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/% b/qa/suites/upgrade/luminous-p2p/% deleted file mode 100644 index e69de29bb2d1..000000000000 diff --git a/qa/suites/upgrade/luminous-p2p/.qa b/qa/suites/upgrade/luminous-p2p/.qa deleted file mode 120000 index a602a0353e75..000000000000 --- a/qa/suites/upgrade/luminous-p2p/.qa +++ /dev/null @@ -1 +0,0 @@ -../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/% b/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/% new file mode 100644 index 000000000000..e69de29bb2d1 diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/.qa b/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/.qa new file mode 120000 index 000000000000..a23f7e045836 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/.qa @@ -0,0 +1 @@ +../../.qa \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/point-to-point-upgrade.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/point-to-point-upgrade.yaml new file mode 100644 index 000000000000..c0b6ebbd5079 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/point-to-point-upgrade.yaml @@ -0,0 +1,226 @@ +meta: +- desc: | + Run ceph on two nodes, using one of them as a client, + with a separate client-only node. + Use xfs beneath the osds. + install ceph/luminous v12.2.2 point version + run workload and upgrade-sequence in parallel + install ceph/luminous v12.2.5 point version + run workload and upgrade-sequence in parallel + install ceph/luminous v12.2.7 point version + run workload and upgrade-sequence in parallel + install ceph/luminous v12.2.8 point version + run workload and upgrade-sequence in parallel + install ceph/luminous v12.2.9 point version + run workload and upgrade-sequence in parallel + install ceph/luminous v12.2.10 point version + run workload and upgrade-sequence in parallel + + install ceph/luminous latest version + run workload and upgrade-sequence in parallel +overrides: + ceph: + log-whitelist: + - reached quota + - scrub + - osd_map_max_advance + - wrongly marked + - FS_DEGRADED + - POOL_APP_NOT_ENABLED + - CACHE_POOL_NO_HIT_SET + - POOL_FULL + - SMALLER_PG + - pool\(s\) full + - OSD_DOWN + - missing hit_sets + - CACHE_POOL_NEAR_FULL + - PG_AVAILABILITY + - PG_DEGRADED + - application not enabled + - overall HEALTH_ + fs: xfs + conf: + mon: + mon debug unsafe allow tier with nonempty snaps: true + mon warn on pool no app: false + osd: + osd map max advance: 1000 + osd_class_load_list: "cephfs hello journal lock log numops rbd refcount + replica_log rgw sdk statelog timeindex user version" + osd_class_default_list: "cephfs hello journal lock log numops rbd refcount + replica_log rgw sdk statelog timeindex user version" + client: + rgw_crypt_require_ssl: false + rgw crypt s3 kms encryption keys: testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= testkey-2=aWIKTWFrZWZpbGUKbWFuCm91dApzcmMKVGVzdGluZwo= +roles: +- - mon.a + - mds.a + - osd.0 + - osd.1 + - osd.2 + - mgr.x +- - mon.b + - mon.c + - osd.3 + - osd.4 + - osd.5 + - client.0 +- - client.1 +openstack: +- volumes: # attached to each instance + count: 3 + size: 30 # GB +tasks: +- print: "**** v12.2.2 about to install" +- install: + tag: v12.2.2 + # line below can be removed its from jewel test + #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev', 'librgw2'] +- print: "**** done v12.2.2 install" +- ceph: + fs: xfs + add_osds_to_crush: true +- print: "**** done ceph xfs" +- sequential: + - workload +- print: "**** done workload v12.2.2" + +#### upgrade to v12.2.5 +- install.upgrade: + #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] + mon.a: + tag: v12.2.5 + mon.b: + tag: v12.2.5 + # Note that client.a IS NOT upgraded at this point +- parallel: + - workload_luminous + - upgrade-sequence_luminous +- print: "**** done parallel luminous v12.2.5" + +#### upgrade to v12.2.7 +- install.upgrade: + #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] + mon.a: + tag: v12.2.7 + mon.b: + tag: v12.2.7 + # Note that client.a IS NOT upgraded at this point +- parallel: + - workload_luminous + - upgrade-sequence_luminous +- print: "**** done parallel luminous v12.2.7" + +#### upgrade to v12.2.8 +- install.upgrade: + #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] + mon.a: + tag: v12.2.8 + mon.b: + tag: v12.2.8 + # Note that client.a IS NOT upgraded at this point +- parallel: + - workload_luminous + - upgrade-sequence_luminous +- print: "**** done parallel luminous v12.2.8" + + +#### upgrade to v12.2.9 +- install.upgrade: + #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] + mon.a: + tag: v12.2.9 + mon.b: + tag: v12.2.9 + # Note that client.a IS NOT upgraded at this point +- parallel: + - workload_luminous + - upgrade-sequence_luminous +- print: "**** done parallel luminous v12.2.9" + +#### upgrade to v12.2.10 +- install.upgrade: + #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] + mon.a: + tag: v12.2.10 + mon.b: + tag: v12.2.10 + # Note that client.a IS NOT upgraded at this point +- parallel: + - workload_luminous + - upgrade-sequence_luminous +- print: "**** done parallel luminous v12.2.10" + + +#### upgrade to latest luminous +- install.upgrade: + #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] + mon.a: + mon.b: + # Note that client.a IS NOT upgraded at this point +- parallel: + - workload_luminous + - upgrade-sequence_luminous +- print: "**** done parallel luminous branch" + +####################### +workload: + sequential: + - workunit: + clients: + client.0: + - suites/blogbench.sh +workload_luminous: + full_sequential: + - workunit: + tag: v12.2.2 + clients: + client.1: + - rados/test.sh + - cls + env: + CLS_RBD_GTEST_FILTER: '*:-TestClsRbd.snapshots_namespaces' + - print: "**** done rados/test.sh & cls workload_luminous" + - sequential: + - rgw: [client.0] + - print: "**** done rgw workload_luminous" + - s3tests: + client.0: + force-branch: ceph-luminous + rgw_server: client.0 + scan_for_encryption_keys: false + - print: "**** done s3tests workload_luminous" +upgrade-sequence_luminous: + sequential: + - print: "**** done branch: luminous install.upgrade" + - ceph.restart: [mds.a] + - sleep: + duration: 60 + - ceph.restart: [osd.0] + - sleep: + duration: 30 + - ceph.restart: [osd.1] + - sleep: + duration: 30 + - ceph.restart: [osd.2] + - sleep: + duration: 30 + - ceph.restart: [osd.3] + - sleep: + duration: 30 + - ceph.restart: [osd.4] + - sleep: + duration: 30 + - ceph.restart: [osd.5] + - sleep: + duration: 60 + - ceph.restart: [mon.a] + - sleep: + duration: 60 + - ceph.restart: [mon.b] + - sleep: + duration: 60 + - ceph.restart: [mon.c] + - sleep: + duration: 60 + - print: "**** done ceph.restart all luminous branch mds/osd/mon" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/supported b/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/supported new file mode 120000 index 000000000000..79010c36a599 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-parallel/supported @@ -0,0 +1 @@ +../../../../distros/supported \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/% b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/% new file mode 100644 index 000000000000..e69de29bb2d1 diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/+ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/+ new file mode 100644 index 000000000000..e69de29bb2d1 diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/.qa b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/.qa new file mode 120000 index 000000000000..a602a0353e75 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/openstack.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/openstack.yaml new file mode 100644 index 000000000000..a0d5c20195be --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/openstack.yaml @@ -0,0 +1,6 @@ +openstack: + - machine: + disk: 100 # GB + - volumes: # attached to each instance + count: 3 + size: 30 # GB diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/start.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/start.yaml new file mode 100644 index 000000000000..4f40219b5267 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/0-cluster/start.yaml @@ -0,0 +1,20 @@ +meta: +- desc: | + Run ceph on two nodes, + with a separate client-only node. + Use xfs beneath the osds. +overrides: + ceph: + fs: xfs +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 +- - osd.3 + - osd.4 + - osd.5 +- - client.0 diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/1-ceph-install/luminous.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/1-ceph-install/luminous.yaml new file mode 100644 index 000000000000..b66e0ca99f06 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/1-ceph-install/luminous.yaml @@ -0,0 +1,19 @@ +meta: +- desc: install ceph/luminous latest +tasks: +- install: + tag: v12.2.10 + exclude_packages: ['librados3'] + extra_packages: ['librados2'] +- print: "**** done install luminous v12.2.10" +- ceph: +- exec: + osd.0: + - ceph osd require-osd-release luminous + - ceph osd set-require-min-compat-client luminous +- print: "**** done ceph" +overrides: + ceph: + conf: + mon: + mon warn on osd down out interval zero: false diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/1.1-pg-log-overrides/normal_pg_log.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/1.1-pg-log-overrides/normal_pg_log.yaml new file mode 100644 index 000000000000..8b137891791f --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/1.1-pg-log-overrides/normal_pg_log.yaml @@ -0,0 +1 @@ + diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/1.1-pg-log-overrides/short_pg_log.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/1.1-pg-log-overrides/short_pg_log.yaml new file mode 100644 index 000000000000..20cc101dea1b --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/1.1-pg-log-overrides/short_pg_log.yaml @@ -0,0 +1,6 @@ +overrides: + ceph: + conf: + global: + osd_min_pg_log_entries: 1 + osd_max_pg_log_entries: 2 diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/2-partial-upgrade/.qa b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/2-partial-upgrade/.qa new file mode 120000 index 000000000000..a602a0353e75 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/2-partial-upgrade/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/2-partial-upgrade/firsthalf.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/2-partial-upgrade/firsthalf.yaml new file mode 100644 index 000000000000..a73b87beb4c5 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/2-partial-upgrade/firsthalf.yaml @@ -0,0 +1,17 @@ +meta: +- desc: | + install upgrade ceph/-x on one node only + 1st half + restart : osd.0,1,2 +tasks: +- install.upgrade: + osd.0: +- print: "**** done install.upgrade osd.0" +- ceph.restart: + daemons: [mon.a,mon.b,mon.c,osd.0, osd.1, osd.2] +- print: "**** done ceph.restart 1st half" +- exec: + osd.0: + - ceph osd set pglog_hardlimit && exit 1 || true + - ceph osd dump --format=json-pretty | grep "flags" +- print: "**** try to set pglog_hardlimit, should not succeed" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/3-thrash/.qa b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/3-thrash/.qa new file mode 120000 index 000000000000..a602a0353e75 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/3-thrash/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/3-thrash/default.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/3-thrash/default.yaml new file mode 100644 index 000000000000..b3fddefc7a35 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/3-thrash/default.yaml @@ -0,0 +1,25 @@ +meta: +- desc: | + randomly kill and revive osd + small chance to increase the number of pgs +overrides: + ceph: + log-whitelist: + - but it is still running + - wrongly marked me down + - objects unfound and apparently lost + - log bound mismatch +tasks: +- parallel: + - stress-tasks +stress-tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgpnum_fix: 1 + chance_thrash_cluster_full: 0 + chance_thrash_pg_upmap: 0 + chance_thrash_pg_upmap_items: 0 + disable_objectstore_tool_tests: true + chance_force_recovery: 0 +- print: "**** done thrashosds 3-thrash" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/+ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/+ new file mode 100644 index 000000000000..e69de29bb2d1 diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/.qa b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/.qa new file mode 120000 index 000000000000..a602a0353e75 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/radosbench.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/radosbench.yaml new file mode 100644 index 000000000000..626ae8ea6653 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/radosbench.yaml @@ -0,0 +1,40 @@ +meta: +- desc: | + run randomized correctness test for rados operations + generate write load with rados bench +stress-tasks: +- full_sequential: + - radosbench: + clients: [client.0] + time: 150 + - radosbench: + clients: [client.0] + time: 150 + - radosbench: + clients: [client.0] + time: 150 + - radosbench: + clients: [client.0] + time: 150 + - radosbench: + clients: [client.0] + time: 150 + - radosbench: + clients: [client.0] + time: 150 + - radosbench: + clients: [client.0] + time: 150 + - radosbench: + clients: [client.0] + time: 150 + - radosbench: + clients: [client.0] + time: 150 + - radosbench: + clients: [client.0] + time: 150 + - radosbench: + clients: [client.0] + time: 150 +- print: "**** done radosbench 7-workload" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/rbd-cls.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/rbd-cls.yaml new file mode 100644 index 000000000000..f8cc4d8ac32b --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/rbd-cls.yaml @@ -0,0 +1,10 @@ +meta: +- desc: | + run basic cls tests for rbd +stress-tasks: +- workunit: + branch: luminous + clients: + client.0: + - cls/test_cls_rbd.sh +- print: "**** done cls/test_cls_rbd.sh 5-workload" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/rbd-import-export.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/rbd-import-export.yaml new file mode 100644 index 000000000000..30a677af65cf --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/rbd-import-export.yaml @@ -0,0 +1,12 @@ +meta: +- desc: | + run basic import/export cli tests for rbd +stress-tasks: +- workunit: + branch: luminous + clients: + client.0: + - rbd/import_export.sh + env: + RBD_CREATE_ARGS: --new-format +- print: "**** done rbd/import_export.sh 5-workload" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/rbd_api.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/rbd_api.yaml new file mode 100644 index 000000000000..9079aa33b897 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/rbd_api.yaml @@ -0,0 +1,10 @@ +meta: +- desc: | + librbd C and C++ api tests +stress-tasks: +- workunit: + branch: luminous + clients: + client.0: + - rbd/test_librbd.sh +- print: "**** done rbd/test_librbd.sh 7-workload" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/readwrite.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/readwrite.yaml new file mode 100644 index 000000000000..41e34d6d7d32 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/readwrite.yaml @@ -0,0 +1,16 @@ +meta: +- desc: | + randomized correctness test for rados operations on a replicated pool, + using only reads, writes, and deletes +stress-tasks: +- full_sequential: + - rados: + clients: [client.0] + ops: 4000 + objects: 500 + write_append_excl: false + op_weights: + read: 45 + write: 45 + delete: 10 +- print: "**** done rados/readwrite 5-workload" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/snaps-few-objects.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/snaps-few-objects.yaml new file mode 100644 index 000000000000..f56d0de0f745 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/4-workload/snaps-few-objects.yaml @@ -0,0 +1,18 @@ +meta: +- desc: | + randomized correctness test for rados operations on a replicated pool with snapshot operations +stress-tasks: +- full_sequential: + - rados: + clients: [client.0] + ops: 4000 + objects: 50 + write_append_excl: false + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 +- print: "**** done rados/snaps-few-objects 5-workload" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/5-finish-upgrade.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/5-finish-upgrade.yaml new file mode 100644 index 000000000000..9d5a96c6cecd --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/5-finish-upgrade.yaml @@ -0,0 +1,14 @@ +tasks: +- install.upgrade: + osd.3: + client.0: +- ceph.restart: + daemons: [osd.3, osd.4, osd.5] + wait-for-healthy: false + wait-for-osds-up: true +- exec: + osd.0: + - ceph osd set pglog_hardlimit + - ceph osd dump --format=json-pretty | grep "flags" +- print: "**** try to set pglog_hardlimit again, should succeed" + diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/+ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/+ new file mode 100644 index 000000000000..e69de29bb2d1 diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/.qa b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/.qa new file mode 120000 index 000000000000..a602a0353e75 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/rbd-python.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/rbd-python.yaml new file mode 100644 index 000000000000..56ba21d7a180 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/rbd-python.yaml @@ -0,0 +1,9 @@ +meta: +- desc: | + librbd python api tests +tasks: +- workunit: + clients: + client.0: + - rbd/test_librbd_python.sh +- print: "**** done rbd/test_librbd_python.sh 9-workload" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/rgw-swift.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/rgw-swift.yaml new file mode 100644 index 000000000000..76e5d6fc2315 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/rgw-swift.yaml @@ -0,0 +1,11 @@ +meta: +- desc: | + swift api tests for rgw +tasks: +- rgw: + client.0: +- print: "**** done rgw 9-workload" +- swift: + client.0: + rgw_server: client.0 +- print: "**** done swift 9-workload" diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/snaps-many-objects.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/snaps-many-objects.yaml new file mode 100644 index 000000000000..805bf97c3792 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/7-final-workload/snaps-many-objects.yaml @@ -0,0 +1,16 @@ +meta: +- desc: | + randomized correctness test for rados operations on a replicated pool with snapshot operations +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 500 + write_append_excl: false + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/supported b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/supported new file mode 120000 index 000000000000..79010c36a599 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/supported @@ -0,0 +1 @@ +../../../../distros/supported \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/thrashosds-health.yaml b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/thrashosds-health.yaml new file mode 120000 index 000000000000..e0426dbe4993 --- /dev/null +++ b/qa/suites/upgrade/luminous-p2p/luminous-p2p-stress-split/thrashosds-health.yaml @@ -0,0 +1 @@ +../../../../tasks/thrashosds-health.yaml \ No newline at end of file diff --git a/qa/suites/upgrade/luminous-p2p/point-to-point-upgrade.yaml b/qa/suites/upgrade/luminous-p2p/point-to-point-upgrade.yaml deleted file mode 100644 index c0b6ebbd5079..000000000000 --- a/qa/suites/upgrade/luminous-p2p/point-to-point-upgrade.yaml +++ /dev/null @@ -1,226 +0,0 @@ -meta: -- desc: | - Run ceph on two nodes, using one of them as a client, - with a separate client-only node. - Use xfs beneath the osds. - install ceph/luminous v12.2.2 point version - run workload and upgrade-sequence in parallel - install ceph/luminous v12.2.5 point version - run workload and upgrade-sequence in parallel - install ceph/luminous v12.2.7 point version - run workload and upgrade-sequence in parallel - install ceph/luminous v12.2.8 point version - run workload and upgrade-sequence in parallel - install ceph/luminous v12.2.9 point version - run workload and upgrade-sequence in parallel - install ceph/luminous v12.2.10 point version - run workload and upgrade-sequence in parallel - - install ceph/luminous latest version - run workload and upgrade-sequence in parallel -overrides: - ceph: - log-whitelist: - - reached quota - - scrub - - osd_map_max_advance - - wrongly marked - - FS_DEGRADED - - POOL_APP_NOT_ENABLED - - CACHE_POOL_NO_HIT_SET - - POOL_FULL - - SMALLER_PG - - pool\(s\) full - - OSD_DOWN - - missing hit_sets - - CACHE_POOL_NEAR_FULL - - PG_AVAILABILITY - - PG_DEGRADED - - application not enabled - - overall HEALTH_ - fs: xfs - conf: - mon: - mon debug unsafe allow tier with nonempty snaps: true - mon warn on pool no app: false - osd: - osd map max advance: 1000 - osd_class_load_list: "cephfs hello journal lock log numops rbd refcount - replica_log rgw sdk statelog timeindex user version" - osd_class_default_list: "cephfs hello journal lock log numops rbd refcount - replica_log rgw sdk statelog timeindex user version" - client: - rgw_crypt_require_ssl: false - rgw crypt s3 kms encryption keys: testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= testkey-2=aWIKTWFrZWZpbGUKbWFuCm91dApzcmMKVGVzdGluZwo= -roles: -- - mon.a - - mds.a - - osd.0 - - osd.1 - - osd.2 - - mgr.x -- - mon.b - - mon.c - - osd.3 - - osd.4 - - osd.5 - - client.0 -- - client.1 -openstack: -- volumes: # attached to each instance - count: 3 - size: 30 # GB -tasks: -- print: "**** v12.2.2 about to install" -- install: - tag: v12.2.2 - # line below can be removed its from jewel test - #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev', 'librgw2'] -- print: "**** done v12.2.2 install" -- ceph: - fs: xfs - add_osds_to_crush: true -- print: "**** done ceph xfs" -- sequential: - - workload -- print: "**** done workload v12.2.2" - -#### upgrade to v12.2.5 -- install.upgrade: - #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - mon.a: - tag: v12.2.5 - mon.b: - tag: v12.2.5 - # Note that client.a IS NOT upgraded at this point -- parallel: - - workload_luminous - - upgrade-sequence_luminous -- print: "**** done parallel luminous v12.2.5" - -#### upgrade to v12.2.7 -- install.upgrade: - #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - mon.a: - tag: v12.2.7 - mon.b: - tag: v12.2.7 - # Note that client.a IS NOT upgraded at this point -- parallel: - - workload_luminous - - upgrade-sequence_luminous -- print: "**** done parallel luminous v12.2.7" - -#### upgrade to v12.2.8 -- install.upgrade: - #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - mon.a: - tag: v12.2.8 - mon.b: - tag: v12.2.8 - # Note that client.a IS NOT upgraded at this point -- parallel: - - workload_luminous - - upgrade-sequence_luminous -- print: "**** done parallel luminous v12.2.8" - - -#### upgrade to v12.2.9 -- install.upgrade: - #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - mon.a: - tag: v12.2.9 - mon.b: - tag: v12.2.9 - # Note that client.a IS NOT upgraded at this point -- parallel: - - workload_luminous - - upgrade-sequence_luminous -- print: "**** done parallel luminous v12.2.9" - -#### upgrade to v12.2.10 -- install.upgrade: - #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - mon.a: - tag: v12.2.10 - mon.b: - tag: v12.2.10 - # Note that client.a IS NOT upgraded at this point -- parallel: - - workload_luminous - - upgrade-sequence_luminous -- print: "**** done parallel luminous v12.2.10" - - -#### upgrade to latest luminous -- install.upgrade: - #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - mon.a: - mon.b: - # Note that client.a IS NOT upgraded at this point -- parallel: - - workload_luminous - - upgrade-sequence_luminous -- print: "**** done parallel luminous branch" - -####################### -workload: - sequential: - - workunit: - clients: - client.0: - - suites/blogbench.sh -workload_luminous: - full_sequential: - - workunit: - tag: v12.2.2 - clients: - client.1: - - rados/test.sh - - cls - env: - CLS_RBD_GTEST_FILTER: '*:-TestClsRbd.snapshots_namespaces' - - print: "**** done rados/test.sh & cls workload_luminous" - - sequential: - - rgw: [client.0] - - print: "**** done rgw workload_luminous" - - s3tests: - client.0: - force-branch: ceph-luminous - rgw_server: client.0 - scan_for_encryption_keys: false - - print: "**** done s3tests workload_luminous" -upgrade-sequence_luminous: - sequential: - - print: "**** done branch: luminous install.upgrade" - - ceph.restart: [mds.a] - - sleep: - duration: 60 - - ceph.restart: [osd.0] - - sleep: - duration: 30 - - ceph.restart: [osd.1] - - sleep: - duration: 30 - - ceph.restart: [osd.2] - - sleep: - duration: 30 - - ceph.restart: [osd.3] - - sleep: - duration: 30 - - ceph.restart: [osd.4] - - sleep: - duration: 30 - - ceph.restart: [osd.5] - - sleep: - duration: 60 - - ceph.restart: [mon.a] - - sleep: - duration: 60 - - ceph.restart: [mon.b] - - sleep: - duration: 60 - - ceph.restart: [mon.c] - - sleep: - duration: 60 - - print: "**** done ceph.restart all luminous branch mds/osd/mon" diff --git a/qa/suites/upgrade/luminous-p2p/supported b/qa/suites/upgrade/luminous-p2p/supported deleted file mode 120000 index dd0d7f1d5bdc..000000000000 --- a/qa/suites/upgrade/luminous-p2p/supported +++ /dev/null @@ -1 +0,0 @@ -../../../distros/supported/ \ No newline at end of file