Use xfs beneath the osds.
install ceph/luminous v12.2.2 point version
run workload and upgrade-sequence in parallel
- install ceph/luminous latest version
+ install ceph/luminous v12.2.5 point version
run workload and upgrade-sequence in parallel
- install ceph/-x version (luminous or master/mimic)
+ install ceph/luminous latest version
run workload and upgrade-sequence in parallel
overrides:
ceph:
- scrub
- osd_map_max_advance
- wrongly marked
+ - FS_DEGRADED
+ - POOL_APP_NOT_ENABLED
+ - CACHE_POOL_NO_HIT_SET
+ - POOL_FULL
+ - SMALLER_PG
+ - pool\(s\) full
+ - OSD_DOWN
+ - missing hit_sets
+ - CACHE_POOL_NEAR_FULL
+ - PG_AVAILABILITY
+ - PG_DEGRADED
+ - application not enabled
fs: xfs
conf:
mon:
- print: "**** done ceph xfs"
- sequential:
- workload
-- print: "**** done workload"
+- print: "**** done workload v12.2.2"
+
+#### upgrade to v12.2.5
- install.upgrade:
#exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev']
mon.a:
- branch: luminous
+ tag: v12.2.5
mon.b:
- branch: luminous
+ tag: v12.2.5
# Note that client.a IS NOT upgraded at this point
- parallel:
- workload_luminous
- upgrade-sequence_luminous
-- print: "**** done parallel luminous branch"
+- print: "**** done parallel luminous v12.2.5"
+#### upgrade to latest luminous
- install.upgrade:
#exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev']
- client.1:
- branch: luminous
-- print: "**** done branch: luminous install.upgrade on client.1"
-- install.upgrade:
mon.a:
+ branch: luminous
mon.b:
-- print: "**** done branch: -x install.upgrade on mon.a and mon.b"
+ branch: luminous
+ # Note that client.a IS NOT upgraded at this point
- parallel:
- - workload_x
- - upgrade-sequence_x
-- print: "**** done parallel -x branch"
-- exec:
- osd.0:
- - ceph osd set-require-min-compat-client luminous
-# Run librados tests on the -x upgraded cluster
-- install.upgrade:
- client.1:
-- workunit:
- branch: luminous
- clients:
- client.1:
- - rados/test.sh
- - cls
-- print: "**** done final test on -x cluster"
+ - workload_luminous
+ - upgrade-sequence_luminous
+- print: "**** done parallel luminous branch"
+
#######################
workload:
sequential:
- sleep:
duration: 60
- print: "**** done ceph.restart all luminous branch mds/osd/mon"
-workload_x:
- sequential:
- - workunit:
- branch: luminous
- clients:
- client.1:
- - rados/test-upgrade-to-mimic.sh
- - cls
- - print: "**** done rados/test-upgrade-to-mimic.sh & cls workload_x NOT upgraded client"
- - workunit:
- branch: luminous
- clients:
- client.0:
- - rados/test-upgrade-to-mimic.sh
- - cls
- - print: "**** done rados/test.sh & cls workload_x upgraded client"
- - rgw: [client.1]
- - print: "**** done rgw workload_x"
- - s3tests:
- client.1:
- force-branch: ceph-luminous
- rgw_server: client.1
- scan_for_encryption_keys: false
- - print: "**** done s3tests workload_x"
-upgrade-sequence_x:
- sequential:
- - ceph.restart: [mds.a]
- - sleep:
- duration: 60
- - ceph.restart: [mon.a]
- - sleep:
- duration: 60
- - ceph.restart: [mon.b]
- - sleep:
- duration: 60
- - ceph.restart: [mon.c]
- - sleep:
- duration: 60
- - ceph.restart: [osd.0]
- - sleep:
- duration: 30
- - ceph.restart: [osd.1]
- - sleep:
- duration: 30
- - ceph.restart: [osd.2]
- - sleep:
- duration: 30
- - ceph.restart: [osd.3]
- - sleep:
- duration: 30
- - ceph.restart: [osd.4]
- - sleep:
- duration: 30
- - ceph.restart:
- daemons: [osd.5]
- wait-for-healthy: false
- wait-for-up-osds: true
- - ceph.restart:
- daemons: [mgr.x]
- wait-for-healthy: false
- - exec:
- osd.0:
- - ceph osd require-osd-release luminous
- - ceph.healthy:
- - print: "**** done ceph.restart all -x branch mds/osd/mon"