+++ /dev/null
-../.qa/
\ No newline at end of file
+++ /dev/null
-../.qa/
\ No newline at end of file
+++ /dev/null
-.qa/cephfs/clusters/3-mds.yaml
\ No newline at end of file
+++ /dev/null
-.qa/cephfs/conf
\ No newline at end of file
+++ /dev/null
-.qa/cephfs/objectstore-ec
\ No newline at end of file
+++ /dev/null
-../.qa/
\ No newline at end of file
+++ /dev/null
-.qa/cephfs/overrides/frag_enable.yaml
\ No newline at end of file
+++ /dev/null
-../.qa/
\ No newline at end of file
+++ /dev/null
-overrides:
- ceph:
- max_mds: 1
+++ /dev/null
-overrides:
- ceph:
- max_mds: 2
+++ /dev/null
-.qa/cephfs/overrides/whitelist_health.yaml
\ No newline at end of file
+++ /dev/null
-overrides:
- ceph:
- log-whitelist:
- - inconsistent rstat on inode
- conf:
- mds:
- mds debug scatterstat: 0
- mds verify scatter: 0
+++ /dev/null
-.qa/cephfs/overrides/whitelist_wrongly_marked_down.yaml
\ No newline at end of file
+++ /dev/null
-../.qa/
\ No newline at end of file
+++ /dev/null
-meta:
-- desc: |
- install ceph/luminous latest
-tasks:
-- install:
- branch: luminous
- exclude_packages:
- - librados3
- - ceph-mgr-dashboard
- - ceph-mgr-diskprediction-local
- - ceph-mgr-diskprediction-cloud
- - ceph-mgr-rook
- - ceph-mgr-ssh
- extra_packages: ['librados2']
-- print: "**** done installing luminous"
-- ceph:
- mon_bind_addrvec: false
- mon_bind_msgr2: false
- log-whitelist:
- - overall HEALTH_
- - \(FS_
- - \(MDS_
- - \(OSD_
- - \(MON_DOWN\)
- - \(CACHE_POOL_
- - \(POOL_
- - \(MGR_DOWN\)
- - \(PG_
- - \(SMALLER_PGP_NUM\)
- - Monitor daemon marked osd
- - Behind on trimming
- - Manager daemon
- conf:
- global:
- mon warn on pool no app: false
- ms bind msgr2: false
-- exec:
- osd.0:
- - ceph osd require-osd-release luminous
- - ceph osd set-require-min-compat-client luminous
-- print: "**** done ceph"
+++ /dev/null
-tasks:
-- ceph-fuse:
-- print: "**** done luminous client"
-- exec:
- mon.a:
- - ceph fs set cephfs allow_new_snaps true --yes-i-really-mean-it
-- workunit:
- timeout: 5m
- cleanup: false
- clients:
- client.0:
- - fs/snap-hierarchy.sh
-- print: "**** done snap hierarchy"
+++ /dev/null
-tasks:
-- mds_pre_upgrade:
-- print: "**** done mds pre-upgrade sequence"
-- install.upgrade:
- mon.a:
- mon.b:
-- print: "**** done install.upgrade both hosts"
-- ceph.stop: [mds.*]
-- ceph.restart:
- daemons: [mon.*, mgr.*, osd.*, mds.*]
- mon-health-to-clog: false
-- print: "**** done ceph.restart"
+++ /dev/null
-tasks:
-- exec:
- mon.a:
- - ceph status
- - ceph fs dump --format=json-pretty
- - ceph fs set cephfs max_mds 2 && exit 1 || true
-- print: "**** confirmed cannot set max_mds=2"
-- exec:
- mon.a:
- - ceph fs set cephfs allow_new_snaps true
+++ /dev/null
-../.qa/
\ No newline at end of file
+++ /dev/null
-tasks:
-- install.upgrade:
- client.0:
-- print: "**** done install.upgrade on client.0"
-- ceph-fuse:
- client.0:
- mounted: false
-- ceph-fuse:
- client.0:
-- print: "**** done remount client"
+++ /dev/null
-tasks:
-- workunit:
- timeout: 5m
- cleanup: false
- env:
- VERIFY: verify
- clients:
- client.0:
- - fs/snap-hierarchy.sh
-- print: "**** done verify snap hierarchy"
+++ /dev/null
-overrides:
- ceph:
- log-whitelist:
- - bad backtrace on inode
-tasks:
-- cephfs_upgrade_snap:
-- print: "**** upgraded snapshot metadata"
-- exec:
- mon.a:
- - ceph fs set cephfs max_mds 2
-- print: "**** increased max_mds=2"
-- sleep:
- duration: 10
-- exec:
- mon.a:
- - ceph fs dump | grep '^max_mds.*2'
+++ /dev/null
-5-client-sanity.yaml
\ No newline at end of file