+++ /dev/null
-../.qa/
\ No newline at end of file
+++ /dev/null
-overrides:
- ceph:
- log-ignorelist:
- - overall HEALTH_
- - \(CACHE_POOL_NEAR_FULL\)
- - \(CACHE_POOL_NO_HIT_SET\)
-tasks:
-- exec:
- client.0:
- - sudo ceph osd pool create cache 4
- - sudo ceph osd tier add rbd cache
- - sudo ceph osd tier cache-mode cache writeback
- - sudo ceph osd tier set-overlay rbd cache
- - sudo ceph osd pool set cache hit_set_type bloom
- - sudo ceph osd pool set cache hit_set_count 8
- - sudo ceph osd pool set cache hit_set_period 60
- - sudo ceph osd pool set cache target_max_objects 250
+++ /dev/null
-overrides:
- ceph:
- log-ignorelist:
- - overall HEALTH_
- - \(CACHE_POOL_NEAR_FULL\)
- - \(CACHE_POOL_NO_HIT_SET\)
-tasks:
-- exec:
- client.0:
- - sudo ceph osd pool create cache 4
- - sudo ceph osd tier add rbd cache
- - sudo ceph osd tier cache-mode cache writeback
- - sudo ceph osd tier set-overlay rbd cache
- - sudo ceph osd pool set cache hit_set_type bloom
- - sudo ceph osd pool set cache hit_set_count 8
- - sudo ceph osd pool set cache hit_set_period 60
- - sudo ceph osd pool set cache target_max_objects 250
+++ /dev/null
-../.qa/
\ No newline at end of file
+++ /dev/null
-overrides:
- ceph:
- log-ignorelist:
- - overall HEALTH_
- - \(CACHE_POOL_NEAR_FULL\)
- - \(CACHE_POOL_NO_HIT_SET\)
-tasks:
-- exec:
- client.0:
- - sudo ceph osd pool create cache 4
- - sudo ceph osd tier add rbd cache
- - sudo ceph osd tier cache-mode cache writeback
- - sudo ceph osd tier set-overlay rbd cache
- - sudo ceph osd pool set cache hit_set_type bloom
- - sudo ceph osd pool set cache hit_set_count 8
- - sudo ceph osd pool set cache hit_set_period 60
- - sudo ceph osd pool set cache target_max_objects 250
+++ /dev/null
-overrides:
- ceph:
- log-ignorelist:
- - overall HEALTH_
- - \(CACHE_POOL_NEAR_FULL\)
- - \(CACHE_POOL_NO_HIT_SET\)
-tasks:
-- exec:
- client.0:
- - sudo ceph osd erasure-code-profile set teuthologyprofile crush-failure-domain=osd m=1 k=2
- - sudo ceph osd pool delete rbd rbd --yes-i-really-really-mean-it
- - sudo ceph osd pool create rbd 4 4 erasure teuthologyprofile
- - sudo ceph osd pool create cache 4
- - sudo ceph osd tier add rbd cache
- - sudo ceph osd tier cache-mode cache writeback
- - sudo ceph osd tier set-overlay rbd cache
- - sudo ceph osd pool set cache hit_set_type bloom
- - sudo ceph osd pool set cache hit_set_count 8
- - sudo ceph osd pool set cache hit_set_period 60
- - sudo ceph osd pool set cache target_max_objects 250
- - rbd pool init rbd
+++ /dev/null
-overrides:
- ceph:
- log-ignorelist:
- - overall HEALTH_
- - \(CACHE_POOL_NEAR_FULL\)
- - \(CACHE_POOL_NO_HIT_SET\)
-tasks:
-- exec:
- client.0:
- - sudo ceph osd pool create cache 4
- - sudo ceph osd tier add rbd cache
- - sudo ceph osd tier cache-mode cache writeback
- - sudo ceph osd tier set-overlay rbd cache
- - sudo ceph osd pool set cache hit_set_type bloom
- - sudo ceph osd pool set cache hit_set_count 8
- - sudo ceph osd pool set cache hit_set_period 60
- - sudo ceph osd pool set cache target_max_objects 250
+++ /dev/null
-overrides:
- ceph:
- log-ignorelist:
- - overall HEALTH_
- - \(CACHE_POOL_NEAR_FULL\)
- - \(CACHE_POOL_NO_HIT_SET\)
-tasks:
-- exec:
- client.0:
- - sudo ceph osd pool create cache 4
- - sudo ceph osd tier add rbd cache
- - sudo ceph osd tier cache-mode cache writeback
- - sudo ceph osd tier set-overlay rbd cache
- - sudo ceph osd pool set cache hit_set_type bloom
- - sudo ceph osd pool set cache hit_set_count 8
- - sudo ceph osd pool set cache hit_set_period 60
- - sudo ceph osd pool set cache target_max_objects 250
+++ /dev/null
-overrides:
- ceph:
- log-ignorelist:
- - overall HEALTH_
- - \(CACHE_POOL_NEAR_FULL\)
- - \(CACHE_POOL_NO_HIT_SET\)
-tasks:
-- exec:
- client.0:
- - sudo ceph osd erasure-code-profile set teuthologyprofile crush-failure-domain=osd m=1 k=2
- - sudo ceph osd pool delete rbd rbd --yes-i-really-really-mean-it
- - sudo ceph osd pool create rbd 4 4 erasure teuthologyprofile
- - sudo ceph osd pool create cache 4
- - sudo ceph osd tier add rbd cache
- - sudo ceph osd tier cache-mode cache writeback
- - sudo ceph osd tier set-overlay rbd cache
- - sudo ceph osd pool set cache hit_set_type bloom
- - sudo ceph osd pool set cache hit_set_count 8
- - sudo ceph osd pool set cache hit_set_period 60
- - sudo ceph osd pool set cache target_max_objects 250
- - rbd pool init rbd
+++ /dev/null
-overrides:
- ceph:
- log-ignorelist:
- - overall HEALTH_
- - \(CACHE_POOL_NEAR_FULL\)
- - \(CACHE_POOL_NO_HIT_SET\)
-tasks:
-- exec:
- client.0:
- - sudo ceph osd pool create cache 4
- - sudo ceph osd tier add rbd cache
- - sudo ceph osd tier cache-mode cache writeback
- - sudo ceph osd tier set-overlay rbd cache
- - sudo ceph osd pool set cache hit_set_type bloom
- - sudo ceph osd pool set cache hit_set_count 8
- - sudo ceph osd pool set cache hit_set_period 60
- - sudo ceph osd pool set cache target_max_objects 250
+++ /dev/null
-overrides:
- ceph:
- log-ignorelist:
- - but it is still running
- - objects unfound and apparently lost
- - overall HEALTH_
- - \(CACHE_POOL_NEAR_FULL\)
- - \(CACHE_POOL_NO_HIT_SET\)
-tasks:
-- exec:
- client.0:
- - sudo ceph osd pool create cache 4
- - sudo ceph osd tier add rbd cache
- - sudo ceph osd tier cache-mode cache writeback
- - sudo ceph osd tier set-overlay rbd cache
- - sudo ceph osd pool set cache hit_set_type bloom
- - sudo ceph osd pool set cache hit_set_count 8
- - sudo ceph osd pool set cache hit_set_period 60
- - sudo ceph osd pool set cache target_max_objects 250
-- thrashosds:
- timeout: 1200