From 3ea9837c8666854a5384bb526c64e63907613f10 Mon Sep 17 00:00:00 2001 From: Yuri Weinstein Date: Fri, 8 Jun 2018 07:53:37 -0700 Subject: [PATCH] qa/tests - added tag 12.2.5. to the mix and removed -x path as unneeded Signed-off-by: Yuri Weinstein --- .../luminous-p2p/point-to-point-upgrade.yaml | 120 ++++-------------- 1 file changed, 28 insertions(+), 92 deletions(-) diff --git a/qa/suites/upgrade/luminous-p2p/point-to-point-upgrade.yaml b/qa/suites/upgrade/luminous-p2p/point-to-point-upgrade.yaml index 0a4ed9ac8125c..f660e2a2f2a6c 100644 --- a/qa/suites/upgrade/luminous-p2p/point-to-point-upgrade.yaml +++ b/qa/suites/upgrade/luminous-p2p/point-to-point-upgrade.yaml @@ -5,9 +5,9 @@ meta: Use xfs beneath the osds. install ceph/luminous v12.2.2 point version run workload and upgrade-sequence in parallel - install ceph/luminous latest version + install ceph/luminous v12.2.5 point version run workload and upgrade-sequence in parallel - install ceph/-x version (luminous or master/mimic) + install ceph/luminous latest version run workload and upgrade-sequence in parallel overrides: ceph: @@ -16,6 +16,18 @@ overrides: - scrub - osd_map_max_advance - wrongly marked + - FS_DEGRADED + - POOL_APP_NOT_ENABLED + - CACHE_POOL_NO_HIT_SET + - POOL_FULL + - SMALLER_PG + - pool\(s\) full + - OSD_DOWN + - missing hit_sets + - CACHE_POOL_NEAR_FULL + - PG_AVAILABILITY + - PG_DEGRADED + - application not enabled fs: xfs conf: mon: @@ -61,44 +73,33 @@ tasks: - print: "**** done ceph xfs" - sequential: - workload -- print: "**** done workload" +- print: "**** done workload v12.2.2" + +#### upgrade to v12.2.5 - install.upgrade: #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] mon.a: - branch: luminous + tag: v12.2.5 mon.b: - branch: luminous + tag: v12.2.5 # Note that client.a IS NOT upgraded at this point - parallel: - workload_luminous - upgrade-sequence_luminous -- print: "**** done parallel luminous branch" +- print: "**** done parallel luminous v12.2.5" +#### upgrade to latest luminous - install.upgrade: #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - client.1: - branch: luminous -- print: "**** done branch: luminous install.upgrade on client.1" -- install.upgrade: mon.a: + branch: luminous mon.b: -- print: "**** done branch: -x install.upgrade on mon.a and mon.b" + branch: luminous + # Note that client.a IS NOT upgraded at this point - parallel: - - workload_x - - upgrade-sequence_x -- print: "**** done parallel -x branch" -- exec: - osd.0: - - ceph osd set-require-min-compat-client luminous -# Run librados tests on the -x upgraded cluster -- install.upgrade: - client.1: -- workunit: - branch: luminous - clients: - client.1: - - rados/test.sh - - cls -- print: "**** done final test on -x cluster" + - workload_luminous + - upgrade-sequence_luminous +- print: "**** done parallel luminous branch" + ####################### workload: sequential: @@ -160,68 +161,3 @@ upgrade-sequence_luminous: - sleep: duration: 60 - print: "**** done ceph.restart all luminous branch mds/osd/mon" -workload_x: - sequential: - - workunit: - branch: luminous - clients: - client.1: - - rados/test-upgrade-to-mimic.sh - - cls - - print: "**** done rados/test-upgrade-to-mimic.sh & cls workload_x NOT upgraded client" - - workunit: - branch: luminous - clients: - client.0: - - rados/test-upgrade-to-mimic.sh - - cls - - print: "**** done rados/test.sh & cls workload_x upgraded client" - - rgw: [client.1] - - print: "**** done rgw workload_x" - - s3tests: - client.1: - force-branch: ceph-luminous - rgw_server: client.1 - scan_for_encryption_keys: false - - print: "**** done s3tests workload_x" -upgrade-sequence_x: - sequential: - - ceph.restart: [mds.a] - - sleep: - duration: 60 - - ceph.restart: [mon.a] - - sleep: - duration: 60 - - ceph.restart: [mon.b] - - sleep: - duration: 60 - - ceph.restart: [mon.c] - - sleep: - duration: 60 - - ceph.restart: [osd.0] - - sleep: - duration: 30 - - ceph.restart: [osd.1] - - sleep: - duration: 30 - - ceph.restart: [osd.2] - - sleep: - duration: 30 - - ceph.restart: [osd.3] - - sleep: - duration: 30 - - ceph.restart: [osd.4] - - sleep: - duration: 30 - - ceph.restart: - daemons: [osd.5] - wait-for-healthy: false - wait-for-up-osds: true - - ceph.restart: - daemons: [mgr.x] - wait-for-healthy: false - - exec: - osd.0: - - ceph osd require-osd-release luminous - - ceph.healthy: - - print: "**** done ceph.restart all -x branch mds/osd/mon" -- 2.39.5