From 657e6c8c3b5bafb96d4179d596de0352479e2a64 Mon Sep 17 00:00:00 2001 From: Guillaume Abrioux Date: Tue, 8 Sep 2020 10:00:06 +0200 Subject: [PATCH] tests: clean legacy clean some legacies since quay.ceph.io migration Signed-off-by: Guillaume Abrioux --- .../ceph-container-common/tasks/registry.yml | 2 +- tests/functional/all_daemons/group_vars/all | 4 ++-- tests/functional/collocation/group_vars/all | 5 ++++- tox-external_clients.ini | 3 --- tox.ini | 20 ++++++------------- 5 files changed, 13 insertions(+), 21 deletions(-) diff --git a/roles/ceph-container-common/tasks/registry.yml b/roles/ceph-container-common/tasks/registry.yml index 4d757ad12..56cb42edc 100644 --- a/roles/ceph-container-common/tasks/registry.yml +++ b/roles/ceph-container-common/tasks/registry.yml @@ -2,4 +2,4 @@ - name: container registry authentication command: '{{ container_binary }} login -u {{ ceph_docker_registry_username }} -p {{ ceph_docker_registry_password }} {{ ceph_docker_registry }}' changed_when: false - no_log: true + no_log: true \ No newline at end of file diff --git a/tests/functional/all_daemons/group_vars/all b/tests/functional/all_daemons/group_vars/all index f20323fb6..863eeb556 100644 --- a/tests/functional/all_daemons/group_vars/all +++ b/tests/functional/all_daemons/group_vars/all @@ -35,7 +35,7 @@ handler_health_osd_check_delay: 10 mds_max_mds: 2 dashboard_admin_password: $sX!cD$rYU6qR^B! grafana_admin_password: +xFRe+RES@7vg24n +ceph_docker_registry: quay.ceph.io node_exporter_container_image: "quay.io/prometheus/node-exporter:v0.17.0" prometheus_container_image: "quay.io/prometheus/prometheus:v2.7.2" -alertmanager_container_image: "quay.io/prometheus/alertmanager:v0.16.2" -ceph_docker_registry: quay.ceph.io +alertmanager_container_image: "quay.io/prometheus/alertmanager:v0.16.2" \ No newline at end of file diff --git a/tests/functional/collocation/group_vars/all b/tests/functional/collocation/group_vars/all index e7fa594ed..7769163f2 100644 --- a/tests/functional/collocation/group_vars/all +++ b/tests/functional/collocation/group_vars/all @@ -20,4 +20,7 @@ handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 dashboard_admin_password: $sX!cD$rYU6qR^B! dashboard_admin_user_ro: true -grafana_admin_password: +xFRe+RES@7vg24n \ No newline at end of file +grafana_admin_password: +xFRe+RES@7vg24n +node_exporter_container_image: "quay.io/prometheus/node-exporter:v0.17.0" +prometheus_container_image: "quay.io/prometheus/prometheus:v2.7.2" +alertmanager_container_image: "quay.io/prometheus/alertmanager:v0.16.2" \ No newline at end of file diff --git a/tox-external_clients.ini b/tox-external_clients.ini index cb93debda..62162e441 100644 --- a/tox-external_clients.ini +++ b/tox-external_clients.ini @@ -57,9 +57,6 @@ commands= fsid=40358a87-ab6e-4bdc-83db-1d909147861c \ external_cluster_mon_ips=192.168.31.10,192.168.31.11,192.168.31.12 \ generate_fsid=false \ - ceph_docker_registry={env:CEPH_DOCKER_REGISTRY:quay.ceph.io} \ - ceph_docker_image=ceph-ci/daemon \ - ceph_docker_image_tag=latest-master \ ceph_dev_branch=master \ ceph_dev_sha1=latest \ ceph_docker_registry_auth=True \ diff --git a/tox.ini b/tox.ini index c568a5863..c580e639c 100644 --- a/tox.ini +++ b/tox.ini @@ -89,9 +89,6 @@ commands= # set up the cluster again ansible-playbook -vv -i {changedir}/{env:INVENTORY} {toxinidir}/{env:PLAYBOOK:site.yml.sample} --extra-vars "\ fetch_directory={env:FETCH_DIRECTORY:{changedir}/fetch} \ - ceph_docker_registry={env:CEPH_DOCKER_REGISTRY:quay.ceph.io} \ - ceph_docker_image={env:CEPH_DOCKER_IMAGE:ceph-ci/daemon} \ - ceph_docker_image_tag={env:CEPH_DOCKER_IMAGE_TAG:latest-master} \ ceph_dev_branch={env:CEPH_DEV_BRANCH:master} \ ceph_dev_sha1={env:CEPH_DEV_SHA1:latest} \ " @@ -109,9 +106,6 @@ commands= ansible-playbook -vv -i {changedir}/{env:INVENTORY} {toxinidir}/infrastructure-playbooks/shrink-osd.yml --extra-vars "\ ireallymeanit=yes \ osd_to_kill={env:OSD_TO_KILL:0} \ - ceph_docker_registry={env:CEPH_DOCKER_REGISTRY:quay.ceph.io} \ - ceph_docker_image={env:CEPH_DOCKER_IMAGE:ceph-ci/daemon} \ - ceph_docker_image_tag={env:CEPH_DOCKER_IMAGE_TAG:latest-master} \ " [shrink-mgr] @@ -147,8 +141,9 @@ commands= ansible-playbook -vv -i {changedir}/{env:INVENTORY} {toxinidir}/infrastructure-playbooks/switch-from-non-containerized-to-containerized-ceph-daemons.yml --extra-vars "\ ireallymeanit=yes \ fetch_directory={env:FETCH_DIRECTORY:{changedir}/fetch} \ - ceph_dev_branch={env:CEPH_DEV_BRANCH:master} \ - ceph_dev_sha1={env:CEPH_DEV_SHA1:latest} \ + ceph_docker_image_tag=latest-master-devel \ + ceph_docker_registry=quay.ceph.io \ + ceph_docker_image=ceph-ci/daemon \ ceph_docker_registry_auth=True \ ceph_docker_registry_username={env:DOCKER_HUB_USERNAME} \ ceph_docker_registry_password={env:DOCKER_HUB_PASSWORD} \ @@ -162,9 +157,6 @@ commands= ansible-playbook -vv -i {changedir}/hosts-2 {toxinidir}/infrastructure-playbooks/add-mon.yml --extra-vars "\ ireallymeanit=yes \ fetch_directory={env:FETCH_DIRECTORY:{changedir}/fetch} \ - ceph_docker_registry={env:CEPH_DOCKER_REGISTRY:quay.ceph.io} \ - ceph_docker_image={env:CEPH_DOCKER_IMAGE:ceph-ci/daemon} \ - ceph_docker_image_tag={env:CEPH_DOCKER_IMAGE_TAG:latest-master} \ ceph_dev_branch={env:CEPH_DEV_BRANCH:master} \ ceph_dev_sha1={env:CEPH_DEV_SHA1:latest} \ " @@ -260,8 +252,8 @@ commands= ceph_docker_registry_username={env:DOCKER_HUB_USERNAME} \ ceph_docker_registry_password={env:DOCKER_HUB_PASSWORD} \ " - ansible-playbook -vv -i {changedir}/hosts {toxinidir}/tests/functional/rgw_multisite.yml --extra-vars "ceph_docker_registry={env:CEPH_DOCKER_REGISTRY:quay.ceph.io} ceph_docker_image={env:CEPH_DOCKER_IMAGE:ceph-ci/daemon} ceph_docker_image_tag={env:CEPH_DOCKER_IMAGE_TAG:latest}" - ansible-playbook --ssh-common-args='-F {changedir}/secondary/vagrant_ssh_config -o ControlMaster=auto -o ControlPersist=600s -o PreferredAuthentications=publickey' -vv -i {changedir}/secondary/hosts {toxinidir}/tests/functional/rgw_multisite.yml --extra-vars "ceph_docker_registry={env:CEPH_DOCKER_REGISTRY:quay.ceph.io} ceph_docker_image={env:CEPH_DOCKER_IMAGE:ceph-ci/daemon} ceph_docker_image_tag={env:CEPH_DOCKER_IMAGE_TAG:latest}" + ansible-playbook -vv -i {changedir}/hosts {toxinidir}/tests/functional/rgw_multisite.yml + ansible-playbook --ssh-common-args='-F {changedir}/secondary/vagrant_ssh_config -o ControlMaster=auto -o ControlPersist=600s -o PreferredAuthentications=publickey' -vv -i {changedir}/secondary/hosts {toxinidir}/tests/functional/rgw_multisite.yml bash -c "cd {changedir}/secondary && vagrant destroy --force" # clean rule after the scenario is complete ansible -i localhost, all -c local -b -m iptables -a 'chain=FORWARD protocol=tcp source=192.168.0.0/16 destination=192.168.0.0/16 jump=ACCEPT action=insert rule_num=1 state=absent' @@ -388,7 +380,7 @@ commands= all_daemons: py.test --reruns 5 --reruns-delay 1 -n 8 --durations=0 --sudo -v --connection=ansible --ansible-inventory={changedir}/{env:INVENTORY} --ssh-config={changedir}/vagrant_ssh_config {toxinidir}/tests/functional/tests # handlers/idempotency test - all_daemons,all_in_one: ansible-playbook -vv -i {changedir}/{env:INVENTORY} {toxinidir}/{env:PLAYBOOK:site.yml.sample} --extra-vars "delegate_facts_host={env:DELEGATE_FACTS_HOST:True} fetch_directory={env:FETCH_DIRECTORY:{changedir}/fetch} ceph_docker_registry={env:CEPH_DOCKER_REGISTRY:quay.ceph.io} ceph_docker_image={env:CEPH_DOCKER_IMAGE:ceph-ci/daemon} ceph_docker_image_tag={env:CEPH_DOCKER_IMAGE_TAG_BIS:latest-bis-master} ceph_dev_branch={env:CEPH_DEV_BRANCH:master} ceph_dev_sha1={env:CEPH_DEV_SHA1:latest}" --extra-vars @ceph-override.json + all_daemons,all_in_one: ansible-playbook -vv -i {changedir}/{env:INVENTORY} {toxinidir}/{env:PLAYBOOK:site.yml.sample} --extra-vars "delegate_facts_host={env:DELEGATE_FACTS_HOST:True} fetch_directory={env:FETCH_DIRECTORY:{changedir}/fetch} ceph_docker_image_tag={env:CEPH_DOCKER_IMAGE_TAG_BIS:latest-bis-master} ceph_dev_branch={env:CEPH_DEV_BRANCH:master} ceph_dev_sha1={env:CEPH_DEV_SHA1:latest}" --extra-vars @ceph-override.json purge: {[purge]commands} switch_to_containers: {[switch-to-containers]commands} -- 2.39.5