]> git.apps.os.sepia.ceph.com Git - ceph.git/commitdiff
tests: Removed 11429 config 17781/head
authorYuri Weinstein <yweinste@redhat.com>
Wed, 21 Sep 2016 16:14:09 +0000 (09:14 -0700)
committerNathan Cutler <ncutler@suse.com>
Mon, 18 Sep 2017 19:28:30 +0000 (21:28 +0200)
Signed-off-by: Yuri Weinstein <yweinste@redhat.com>
(cherry picked from fde4d3bb96abc0c0ca5bca213a0a2423f95e91a2)

Conflicts:
    Cherry-picked manually because, nowadays, we have the entire ceph-qa-suite
    under qa/ in ceph/ceph.git

qa/suites/rados/singleton-nomsgr/all/11429.yaml [deleted file]

diff --git a/qa/suites/rados/singleton-nomsgr/all/11429.yaml b/qa/suites/rados/singleton-nomsgr/all/11429.yaml
deleted file mode 100644 (file)
index 53628e5..0000000
+++ /dev/null
@@ -1,136 +0,0 @@
-# we don't have el7 packages for old releases
-# http://tracker.ceph.com/issues/15139
-os_type: ubuntu
-os_version: "14.04"
-overrides:
-  ceph:
-    conf:
-      mon:
-        debug mon: 20
-        debug ms: 1
-        debug paxos: 20
-        mon warn on legacy crush tunables: false
-        mon min osdmap epochs: 3
-      osd:
-        osd map cache size: 2
-        osd map max advance: 1
-        debug filestore: 20
-        debug journal: 20
-        debug ms: 1
-        debug osd: 20
-    log-whitelist:
-    - osd_map_cache_size
-    - slow request
-    - scrub mismatch
-    - ScrubResult
-    - failed to encode
-roles:
-- - mon.a
-  - osd.0
-  - osd.1
-  - mon.b
-  - mon.c
-  - osd.2
-  - client.0
-openstack:
-  - volumes: # attached to each instance
-      count: 3
-      size: 10 # GB
-tasks:
-- install:
-    branch: v0.80.8
-- print: '**** done installing firefly'
-- ceph:
-    fs: xfs
-- print: '**** done ceph'
-- full_sequential:
-  - ceph_manager.create_pool:
-      args: ['toremove']
-      kwargs:
-        pg_num: 4096
-  - sleep:
-      duration: 30
-  - ceph_manager.wait_for_clean: null
-  - radosbench:
-      clients: [client.0]
-      time: 120
-      size: 1
-      pool: toremove
-      create_pool: false
-      cleanup: false
-  - ceph_manager.remove_pool:
-      args: ['toremove']
-  - exec:
-      client.0:
-        - rados -p rbd ls -
-  - exec:
-      osd.0:
-        - ceph daemon osd.0 config set filestore_blackhole true
-  - ceph.restart:
-      daemons:
-        - osd.0
-        - osd.1
-        - osd.2
-  - exec:
-      osd.0:
-        - sudo grep -c 'load_pgs. skipping PG' /var/log/ceph/ceph-osd.0.log
-  - ceph_manager.wait_for_clean: null
-  - ceph_manager.create_pool:
-      args: ['newpool']
-  - loop:
-      count: 100
-      body:
-        - ceph_manager.set_pool_property:
-            args: ['newpool', 'min_size', 2]
-        - ceph_manager.set_pool_property:
-            args: ['newpool', 'min_size', 1]
-  - ceph_manager.wait_for_clean: null
-  - loop:
-      count: 100
-      body:
-        - ceph_manager.set_pool_property:
-            args: ['newpool', 'min_size', 2]
-        - ceph_manager.set_pool_property:
-            args: ['newpool', 'min_size', 1]
-  - sleep:
-      duration: 10
-  - ceph_manager.wait_for_clean: null
-  - print: '**** done creating zombie pgs'
-
-  - install.upgrade:
-      mon.a:
-        branch: hammer
-  - ceph.restart: [mon.a, mon.b, mon.c, osd.0, osd.1, osd.2]
-  - ceph_manager.wait_for_clean: null
-  - ceph.restart: [osd.0, osd.1, osd.2]
-  - ceph_manager.wait_for_clean: null
-  - exec:
-      osd.0:
-        - sudo grep -c 'Skipping the pg for now' /var/log/ceph/ceph-osd.0.log
-  - print: '**** done verifying hammer upgrade'
-
-  - install.upgrade:
-      mon.a: null
-  - ceph.restart:
-      daemons: [mon.a, mon.b, mon.c, osd.0, osd.1, osd.2]
-      wait-for-healthy: false
-      wait-for-osds-up: true
-  - exec:
-      osd.0:
-        - sleep 300 # http://tracker.ceph.com/issues/17808
-        - ceph osd set require_jewel_osds
-        - ceph osd set sortbitwise
-  - ceph.healthy:
-  - ceph_manager.wait_for_clean: null
-  - ceph.restart: [osd.0, osd.1, osd.2]
-  - ceph_manager.wait_for_clean: null
-  - exec:
-      osd.0:
-        - sudo grep -c 'unable to peek at' /var/log/ceph/ceph-osd.0.log
-  - radosbench:
-      clients: [client.0]
-      time: 5
-      size: 1
-  - ceph.restart: [osd.0, osd.1, osd.2]
-  - ceph_manager.wait_for_clean: null
-  - print: '**** done verifying final upgrade'