From 96e29adf72804cb4fca62c04416b8fe2d23587bb Mon Sep 17 00:00:00 2001 From: Alfredo Deza Date: Mon, 24 Sep 2018 14:44:31 -0400 Subject: [PATCH] ceph-volume tests/functional update dmcrypt test playbooks to use --cluster Signed-off-by: Alfredo Deza --- .../lvm/centos7/bluestore/dmcrypt/test.yml | 18 +++++++++--------- .../lvm/centos7/filestore/dmcrypt/test.yml | 16 ++++++++-------- .../lvm/xenial/bluestore/dmcrypt/test.yml | 18 +++++++++--------- .../lvm/xenial/filestore/dmcrypt/test.yml | 16 ++++++++-------- 4 files changed, 34 insertions(+), 34 deletions(-) diff --git a/src/ceph-volume/ceph_volume/tests/functional/lvm/centos7/bluestore/dmcrypt/test.yml b/src/ceph-volume/ceph_volume/tests/functional/lvm/centos7/bluestore/dmcrypt/test.yml index d882293def721..bebe6dc36ba51 100644 --- a/src/ceph-volume/ceph_volume/tests/functional/lvm/centos7/bluestore/dmcrypt/test.yml +++ b/src/ceph-volume/ceph_volume/tests/functional/lvm/centos7/bluestore/dmcrypt/test.yml @@ -17,10 +17,10 @@ tasks: - name: destroy osd.2 - command: "ceph osd destroy osd.2 --yes-i-really-mean-it" + command: "ceph --cluster {{ cluster }} osd destroy osd.2 --yes-i-really-mean-it" - name: destroy osd.0 - command: "ceph osd destroy osd.0 --yes-i-really-mean-it" + command: "ceph --cluster {{ cluster }} osd destroy osd.0 --yes-i-really-mean-it" - hosts: osds become: yes @@ -28,23 +28,23 @@ # osd.2 device - name: zap /dev/sdd1 - command: "ceph-volume lvm zap /dev/sdd1 --destroy" + command: "ceph-volume --cluster {{ cluster }} lvm zap /dev/sdd1 --destroy" environment: CEPH_VOLUME_DEBUG: 1 - name: redeploy osd.2 using /dev/sdd1 - command: "ceph-volume lvm create --bluestore --data /dev/sdd1 --osd-id 2" + command: "ceph-volume --cluster {{ cluster }} lvm create --bluestore --data /dev/sdd1 --osd-id 2" environment: CEPH_VOLUME_DEBUG: 1 # osd.0 lv - name: zap test_group/data-lv1 - command: "ceph-volume lvm zap test_group/data-lv1" + command: "ceph-volume --cluster {{ cluster }} lvm zap test_group/data-lv1" environment: CEPH_VOLUME_DEBUG: 1 - name: redeploy osd.0 using test_group/data-lv1 - command: "ceph-volume lvm create --bluestore --data test_group/data-lv1 --osd-id 0" + command: "ceph-volume --cluster {{ cluster }} lvm create --bluestore --data test_group/data-lv1 --osd-id 0" environment: CEPH_VOLUME_DEBUG: 1 @@ -59,7 +59,7 @@ tasks: - name: destroy osd.0 - command: "ceph osd destroy osd.0 --yes-i-really-mean-it" + command: "ceph --cluster {{ cluster }} osd destroy osd.0 --yes-i-really-mean-it" - hosts: osds @@ -68,12 +68,12 @@ - name: zap test_group/data-lv1 - command: "ceph-volume lvm zap test_group/data-lv1" + command: "ceph-volume --cluster {{ cluster }} lvm zap test_group/data-lv1" environment: CEPH_VOLUME_DEBUG: 1 - name: prepare osd.0 using test_group/data-lv1 - command: "ceph-volume lvm prepare --bluestore --data test_group/data-lv1 --osd-id 0" + command: "ceph-volume --cluster {{ cluster }} lvm prepare --bluestore --data test_group/data-lv1 --osd-id 0" environment: CEPH_VOLUME_DEBUG: 1 diff --git a/src/ceph-volume/ceph_volume/tests/functional/lvm/centos7/filestore/dmcrypt/test.yml b/src/ceph-volume/ceph_volume/tests/functional/lvm/centos7/filestore/dmcrypt/test.yml index 5dc67ade18c16..c48e4becece7d 100644 --- a/src/ceph-volume/ceph_volume/tests/functional/lvm/centos7/filestore/dmcrypt/test.yml +++ b/src/ceph-volume/ceph_volume/tests/functional/lvm/centos7/filestore/dmcrypt/test.yml @@ -19,10 +19,10 @@ tasks: - name: destroy osd.2 - command: "ceph osd destroy osd.2 --yes-i-really-mean-it" + command: "ceph --cluster {{ cluster }} osd destroy osd.2 --yes-i-really-mean-it" - name: destroy osd.0 - command: "ceph osd destroy osd.0 --yes-i-really-mean-it" + command: "ceph --cluster {{ cluster }} osd destroy osd.0 --yes-i-really-mean-it" - hosts: osds @@ -31,33 +31,33 @@ # osd.2 device - name: zap /dev/sdd1 - command: "ceph-volume lvm zap /dev/sdd1 --destroy" + command: "ceph-volume --cluster {{ cluster }} lvm zap /dev/sdd1 --destroy" environment: CEPH_VOLUME_DEBUG: 1 - name: zap /dev/sdd2 - command: "ceph-volume lvm zap /dev/sdd2 --destroy" + command: "ceph-volume --cluster {{ cluster }} lvm zap /dev/sdd2 --destroy" environment: CEPH_VOLUME_DEBUG: 1 - name: redeploy osd.2 using /dev/sdd1 - command: "ceph-volume lvm create --filestore --data /dev/sdd1 --journal /dev/sdd2 --osd-id 2" + command: "ceph-volume --cluster {{ cluster }} lvm create --filestore --data /dev/sdd1 --journal /dev/sdd2 --osd-id 2" environment: CEPH_VOLUME_DEBUG: 1 # osd.0 lv - name: zap test_group/data-lv1 - command: "ceph-volume lvm zap test_group/data-lv1" + command: "ceph-volume --cluster {{ cluster }} lvm zap test_group/data-lv1" environment: CEPH_VOLUME_DEBUG: 1 - name: zap /dev/sdc1 - command: "ceph-volume lvm zap /dev/sdc1 --destroy" + command: "ceph-volume --cluster {{ cluster }} lvm zap /dev/sdc1 --destroy" environment: CEPH_VOLUME_DEBUG: 1 - name: prepare osd.0 again using test_group/data-lv1 - command: "ceph-volume lvm prepare --filestore --data test_group/data-lv1 --journal /dev/sdc1 --osd-id 0" + command: "ceph-volume --cluster {{ cluster }} lvm prepare --filestore --data test_group/data-lv1 --journal /dev/sdc1 --osd-id 0" environment: CEPH_VOLUME_DEBUG: 1 diff --git a/src/ceph-volume/ceph_volume/tests/functional/lvm/xenial/bluestore/dmcrypt/test.yml b/src/ceph-volume/ceph_volume/tests/functional/lvm/xenial/bluestore/dmcrypt/test.yml index b6db0ac2f90fc..19209b1d21136 100644 --- a/src/ceph-volume/ceph_volume/tests/functional/lvm/xenial/bluestore/dmcrypt/test.yml +++ b/src/ceph-volume/ceph_volume/tests/functional/lvm/xenial/bluestore/dmcrypt/test.yml @@ -17,10 +17,10 @@ tasks: - name: destroy osd.2 - command: "ceph osd destroy osd.2 --yes-i-really-mean-it" + command: "ceph --cluster {{ cluster }} osd destroy osd.2 --yes-i-really-mean-it" - name: destroy osd.0 - command: "ceph osd destroy osd.0 --yes-i-really-mean-it" + command: "ceph --cluster {{ cluster }} osd destroy osd.0 --yes-i-really-mean-it" - hosts: osds @@ -29,23 +29,23 @@ # osd.2 device - name: zap /dev/sdd1 - command: "ceph-volume lvm zap /dev/sdd1 --destroy" + command: "ceph-volume --cluster {{ cluster }} lvm zap /dev/sdd1 --destroy" environment: CEPH_VOLUME_DEBUG: 1 - name: redeploy osd.2 using /dev/sdd1 - command: "ceph-volume lvm create --bluestore --data /dev/sdd1 --osd-id 2" + command: "ceph-volume --cluster {{ cluster }} lvm create --bluestore --data /dev/sdd1 --osd-id 2" environment: CEPH_VOLUME_DEBUG: 1 # osd.0 lv - name: zap test_group/data-lv1 - command: "ceph-volume lvm zap test_group/data-lv1" + command: "ceph-volume --cluster {{ cluster }} lvm zap test_group/data-lv1" environment: CEPH_VOLUME_DEBUG: 1 - name: redeploy osd.0 using test_group/data-lv1 - command: "ceph-volume lvm create --bluestore --data test_group/data-lv1 --osd-id 0" + command: "ceph-volume --cluster {{ cluster }} lvm create --bluestore --data test_group/data-lv1 --osd-id 0" environment: CEPH_VOLUME_DEBUG: 1 @@ -60,7 +60,7 @@ tasks: - name: destroy osd.0 - command: "ceph osd destroy osd.0 --yes-i-really-mean-it" + command: "ceph --cluster {{ cluster }} osd destroy osd.0 --yes-i-really-mean-it" - hosts: osds @@ -68,12 +68,12 @@ tasks: - name: zap test_group/data-lv1 - command: "ceph-volume lvm zap test_group/data-lv1" + command: "ceph-volume --cluster {{ cluster }} lvm zap test_group/data-lv1" environment: CEPH_VOLUME_DEBUG: 1 - name: prepare osd.0 using test_group/data-lv1 - command: "ceph-volume lvm prepare --bluestore --data test_group/data-lv1 --osd-id 0" + command: "ceph-volume --cluster {{ cluster }} lvm prepare --bluestore --data test_group/data-lv1 --osd-id 0" environment: CEPH_VOLUME_DEBUG: 1 diff --git a/src/ceph-volume/ceph_volume/tests/functional/lvm/xenial/filestore/dmcrypt/test.yml b/src/ceph-volume/ceph_volume/tests/functional/lvm/xenial/filestore/dmcrypt/test.yml index 5dc67ade18c16..c48e4becece7d 100644 --- a/src/ceph-volume/ceph_volume/tests/functional/lvm/xenial/filestore/dmcrypt/test.yml +++ b/src/ceph-volume/ceph_volume/tests/functional/lvm/xenial/filestore/dmcrypt/test.yml @@ -19,10 +19,10 @@ tasks: - name: destroy osd.2 - command: "ceph osd destroy osd.2 --yes-i-really-mean-it" + command: "ceph --cluster {{ cluster }} osd destroy osd.2 --yes-i-really-mean-it" - name: destroy osd.0 - command: "ceph osd destroy osd.0 --yes-i-really-mean-it" + command: "ceph --cluster {{ cluster }} osd destroy osd.0 --yes-i-really-mean-it" - hosts: osds @@ -31,33 +31,33 @@ # osd.2 device - name: zap /dev/sdd1 - command: "ceph-volume lvm zap /dev/sdd1 --destroy" + command: "ceph-volume --cluster {{ cluster }} lvm zap /dev/sdd1 --destroy" environment: CEPH_VOLUME_DEBUG: 1 - name: zap /dev/sdd2 - command: "ceph-volume lvm zap /dev/sdd2 --destroy" + command: "ceph-volume --cluster {{ cluster }} lvm zap /dev/sdd2 --destroy" environment: CEPH_VOLUME_DEBUG: 1 - name: redeploy osd.2 using /dev/sdd1 - command: "ceph-volume lvm create --filestore --data /dev/sdd1 --journal /dev/sdd2 --osd-id 2" + command: "ceph-volume --cluster {{ cluster }} lvm create --filestore --data /dev/sdd1 --journal /dev/sdd2 --osd-id 2" environment: CEPH_VOLUME_DEBUG: 1 # osd.0 lv - name: zap test_group/data-lv1 - command: "ceph-volume lvm zap test_group/data-lv1" + command: "ceph-volume --cluster {{ cluster }} lvm zap test_group/data-lv1" environment: CEPH_VOLUME_DEBUG: 1 - name: zap /dev/sdc1 - command: "ceph-volume lvm zap /dev/sdc1 --destroy" + command: "ceph-volume --cluster {{ cluster }} lvm zap /dev/sdc1 --destroy" environment: CEPH_VOLUME_DEBUG: 1 - name: prepare osd.0 again using test_group/data-lv1 - command: "ceph-volume lvm prepare --filestore --data test_group/data-lv1 --journal /dev/sdc1 --osd-id 0" + command: "ceph-volume --cluster {{ cluster }} lvm prepare --filestore --data test_group/data-lv1 --journal /dev/sdc1 --osd-id 0" environment: CEPH_VOLUME_DEBUG: 1 -- 2.39.5