From 5b6e39617fec33fbe1658e19a0833759a7682d45 Mon Sep 17 00:00:00 2001 From: Alfredo Deza Date: Tue, 13 Mar 2018 06:57:57 -0400 Subject: [PATCH] ceph-volume tests use granular env vars for vagrant Fixes the error that would inject an extra argument to vagrant when the environment variable is not present Signed-off-by: Alfredo Deza --- src/ceph-volume/ceph_volume/tests/functional/lvm/tox.ini | 6 +++--- src/ceph-volume/ceph_volume/tests/functional/simple/tox.ini | 6 +++--- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/src/ceph-volume/ceph_volume/tests/functional/lvm/tox.ini b/src/ceph-volume/ceph_volume/tests/functional/lvm/tox.ini index f5d44e68556cc..683ef840a8bc6 100644 --- a/src/ceph-volume/ceph_volume/tests/functional/lvm/tox.ini +++ b/src/ceph-volume/ceph_volume/tests/functional/lvm/tox.ini @@ -39,7 +39,7 @@ changedir= commands= git clone -b {env:CEPH_ANSIBLE_BRANCH:master} --single-branch https://github.com/ceph/ceph-ansible.git {envdir}/tmp/ceph-ansible - vagrant up {env:VAGRANT_GLOBAL_EXTRA_FLAGS:""} --no-provision {posargs:--provider=virtualbox} + vagrant up {env:VAGRANT_UP_FLAGS:"--no-provision"} {posargs:--provider=virtualbox} bash {toxinidir}/../scripts/generate_ssh_config.sh {changedir} # create logical volumes to test with on the vms @@ -58,7 +58,7 @@ commands= testinfra -n 4 --sudo -v --connection=ansible --ansible-inventory={changedir}/hosts {envdir}/tmp/ceph-ansible/tests/functional/tests # reboot all vms - vagrant reload {env:VAGRANT_GLOBAL_EXTRA_FLAGS:""} --no-provision + vagrant reload {env:VAGRANT_RELOAD_FLAGS:"--no-provision"} # retest to ensure cluster came back up correctly after rebooting testinfra -n 4 --sudo -v --connection=ansible --ansible-inventory={changedir}/hosts {envdir}/tmp/ceph-ansible/tests/functional/tests @@ -69,4 +69,4 @@ commands= # retest to ensure cluster came back up correctly testinfra -n 4 --sudo -v --connection=ansible --ansible-inventory={changedir}/hosts {envdir}/tmp/ceph-ansible/tests/functional/tests - vagrant destroy {env:VAGRANT_GLOBAL_EXTRA_FLAGS:""} --force + vagrant destroy {env:VAGRANT_DESTROY_FLAGS:"--force"} diff --git a/src/ceph-volume/ceph_volume/tests/functional/simple/tox.ini b/src/ceph-volume/ceph_volume/tests/functional/simple/tox.ini index 21d9c3a709fa1..bfb265fe987aa 100644 --- a/src/ceph-volume/ceph_volume/tests/functional/simple/tox.ini +++ b/src/ceph-volume/ceph_volume/tests/functional/simple/tox.ini @@ -36,7 +36,7 @@ changedir= commands= git clone -b {env:CEPH_ANSIBLE_BRANCH:master} --single-branch https://github.com/ceph/ceph-ansible.git {envdir}/tmp/ceph-ansible - vagrant up {env:VAGRANT_GLOBAL_EXTRA_FLAGS:""} --no-provision {posargs:--provider=virtualbox} + vagrant up {env:VAGRANT_UP_FLAGS:"--no-provision"} {posargs:--provider=virtualbox} bash {toxinidir}/../scripts/generate_ssh_config.sh {changedir} # use ceph-ansible to deploy a ceph cluster on the vms @@ -52,7 +52,7 @@ commands= ansible-playbook -vv -i {changedir}/hosts {changedir}/test.yml # reboot all vms - vagrant reload {env:VAGRANT_GLOBAL_EXTRA_FLAGS:""} --no-provision + vagrant reload {env:VAGRANT_RELOAD_FLAGS:"--no-provision"} # wait 2 minutes for services to be ready sleep 120 @@ -60,4 +60,4 @@ commands= # retest to ensure cluster came back up correctly after rebooting testinfra -n 4 --sudo -v --connection=ansible --ansible-inventory={changedir}/hosts {envdir}/tmp/ceph-ansible/tests/functional/tests - vagrant destroy {env:VAGRANT_GLOBAL_EXTRA_FLAGS:""} --force + vagrant destroy {env:VAGRANT_DESTROY_FLAGS:"--force"} -- 2.39.5