]> git.apps.os.sepia.ceph.com Git - ceph.git/commitdiff
Fixes #12068
authorYuri Weinstein <yweinste@redhat.com>
Thu, 23 Jul 2015 16:13:11 +0000 (09:13 -0700)
committerYuri Weinstein <yweinste@redhat.com>
Thu, 23 Jul 2015 16:57:58 +0000 (09:57 -0700)
point-to-point test moved from upgrade/hammer and modified to run as -x type upgrade

Signed-off-by: Yuri Weinstein <yweinste@redhat.com>
suites/upgrade/hammer-x/point-to-point-x/% [new file with mode: 0644]
suites/upgrade/hammer-x/point-to-point-x/distros [new symlink]
suites/upgrade/hammer-x/point-to-point-x/point-to-point.yaml [new file with mode: 0644]

diff --git a/suites/upgrade/hammer-x/point-to-point-x/% b/suites/upgrade/hammer-x/point-to-point-x/%
new file mode 100644 (file)
index 0000000..e69de29
diff --git a/suites/upgrade/hammer-x/point-to-point-x/distros b/suites/upgrade/hammer-x/point-to-point-x/distros
new file mode 120000 (symlink)
index 0000000..ca99fee
--- /dev/null
@@ -0,0 +1 @@
+../../../../distros/supported/
\ No newline at end of file
diff --git a/suites/upgrade/hammer-x/point-to-point-x/point-to-point.yaml b/suites/upgrade/hammer-x/point-to-point-x/point-to-point.yaml
new file mode 100644 (file)
index 0000000..07f2c1f
--- /dev/null
@@ -0,0 +1,241 @@
+overrides:
+  ceph:
+    log-whitelist:
+    - scrub
+    - osd_map_max_advance
+    fs: xfs
+    conf:
+      mon:
+        mon debug unsafe allow tier with nonempty snaps: true
+      osd:
+        osd map max advance: 1000
+roles:
+- - mon.a
+  - mds.a
+  - osd.0
+  - osd.1
+  - osd.2
+- - mon.b
+  - mon.c
+  - osd.3
+  - osd.4
+  - osd.5
+  - client.0
+- - client.1
+tasks:
+- print: "****  v0.94 about to install"
+- install:
+    tag: v0.94
+- print: "**** done v0.94 install"
+- ceph:
+   fs: xfs
+- print: "**** done ceph xfs"
+- sequential:
+   - workload
+- print: "**** done workload v0.94"
+- parallel:
+   - workload1
+   - upgrade-sequence1
+- print: "**** done parallel v0.94.2"
+- parallel:
+   - workload_hammer
+   - upgrade-sequence_hammer
+- print: "**** done parallel hammer branch"
+- parallel:
+   - workload_x
+   - upgrade-sequence_x
+- print: "**** done parallel -x branch"
+# Run test.sh on the -x upgraded cluster
+- workunit:
+    clients:
+      client.1:
+      - rados/test.sh
+      - cls
+- print: "**** done final test on -x cluster"
+#######################
+workload:
+   sequential:
+   - workunit:
+       clients:
+         client.0:
+           - suites/blogbench.sh
+   - print: "**** done suites/blogbench.sh workload"
+workload1:
+   sequential:
+   - workunit:
+       clients:
+         client.0:
+         - rados/load-gen-big.sh
+   - print: "**** done rados/load-gen-big.sh workload1"
+   - workunit:
+       clients:
+         client.0:
+         - rados/test.sh
+         - cls
+   - print: "**** done rados/test.sh &  cls workload1"
+   - workunit:
+       clients:
+         client.0:
+         - rbd/test_librbd.sh
+   - print: "**** done rbd/test_librbd.sh workload1"
+upgrade-sequence1:
+   sequential:
+   - install.upgrade:
+       mon.a:
+         tag: v0.94.2
+       mon.b:
+         tag: v0.94.2
+       client.1:
+         tag: v0.94.2
+   - print: "**** done v0.94.2 install.upgrade"
+   - ceph.restart: [mon.a]
+   - sleep:
+       duration: 60
+   - ceph.restart: [mon.b]
+   - sleep:
+       duration: 60
+   - ceph.restart: [mon.c]
+   - sleep:
+       duration: 60
+   - ceph.restart: [mds.a]
+   - sleep:
+       duration: 60
+   - ceph.restart: [osd.0]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.1]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.2]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.3]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.4]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.5]
+   - sleep:
+       duration: 30
+   - print: "**** done ceph.restart all mon/mds/osd v0.94.2"
+workload_hammer:
+   sequential:
+   - workunit:
+       clients:
+         client.1:
+         - rados/test.sh
+         - cls
+   - print: "**** done rados/test.sh &  cls workload_hammer"
+   - rgw: [client.0]
+   - print: "**** done rgw workload_hammer"
+   - s3tests:
+       client.0:
+         force-branch: hammer
+         rgw_server: client.0
+   - print: "**** done s3tests workload_hammer"
+upgrade-sequence_hammer:
+   sequential:
+   - install.upgrade:
+       mon.a:
+         branch: hammer
+       mon.b:
+         branch: hammer
+       # Note that client.a IS NOT upgraded at this point
+       #client.1:
+         #branch: hammer
+   - print: "**** done branch: hammer install.upgrade"
+   - ceph.restart: [mds.a]
+   - sleep:
+       duration: 60
+   - ceph.restart: [osd.0]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.1]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.2]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.3]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.4]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.5]
+   - sleep:
+       duration: 60
+   - ceph.restart: [mon.a]
+   - sleep:
+       duration: 60
+   - ceph.restart: [mon.b]
+   - sleep:
+       duration: 60
+   - ceph.restart: [mon.c]
+   - sleep:
+       duration: 60
+   - print: "**** done ceph.restart all hammer branch mds/osd/mon"
+workload_x:
+   sequential:
+   - workunit:
+       clients:
+         client.1:
+         - rados/test.sh
+         - cls
+   - print: "**** done rados/test.sh &  cls workload_x NOT upgraded  client"
+   - workunit:
+       clients:
+         client.0:
+         - rados/test.sh
+         - cls
+   - print: "**** done rados/test.sh &  cls workload_x upgraded client"
+   - rgw: [client.0]
+   - print: "**** done rgw workload_x"
+   - s3tests:
+       client.0:
+         force-branch: hammer
+         rgw_server: client.0
+   - print: "**** done s3tests workload_x"
+upgrade-sequence_x:
+   sequential:
+   - install.upgrade:
+       mon.a:
+         #branch: hammer
+       mon.b:
+         #branch: hammer
+       # Note that client.a IS NOT upgraded here 
+       #client.1:
+         #branch: hammer
+   - print: "**** done branch: -x install.upgrade"
+   - ceph.restart: [mds.a]
+   - sleep:
+       duration: 60
+   - ceph.restart: [osd.0]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.1]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.2]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.3]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.4]
+   - sleep:
+       duration: 30
+   - ceph.restart: [osd.5]
+   - sleep:
+       duration: 60
+   - ceph.restart: [mon.a]
+   - sleep:
+       duration: 60
+   - ceph.restart: [mon.b]
+   - sleep:
+       duration: 60
+   - ceph.restart: [mon.c]
+   - sleep:
+       duration: 60
+   - print: "**** done ceph.restart all -x branch mds/osd/mon"