]> git.apps.os.sepia.ceph.com Git - ceph-ansible.git/commitdiff
Tests: Add bluestore scenarios
authorGuillaume Abrioux <gabrioux@redhat.com>
Wed, 12 Jul 2017 13:02:32 +0000 (15:02 +0200)
committerGuillaume Abrioux <gabrioux@redhat.com>
Wed, 12 Jul 2017 13:02:32 +0000 (15:02 +0200)
Since we started testing against Luminous, we need to add more scenarios
testing.

Signed-off-by: Guillaume Abrioux <gabrioux@redhat.com>
36 files changed:
Vagrantfile
group_vars/osds.yml.sample
roles/ceph-osd/defaults/main.yml
roles/ceph-osd/tasks/scenarios/dmcrypt-dedicated-journal.yml
tests/functional/centos/7/bs-dmcrypt-dedicated-journal/Vagrantfile [new symlink]
tests/functional/centos/7/bs-dmcrypt-dedicated-journal/group_vars/all [new file with mode: 0644]
tests/functional/centos/7/bs-dmcrypt-dedicated-journal/group_vars/osds [new file with mode: 0644]
tests/functional/centos/7/bs-dmcrypt-dedicated-journal/hosts [new file with mode: 0644]
tests/functional/centos/7/bs-dmcrypt-dedicated-journal/vagrant_variables.yml [new file with mode: 0644]
tests/functional/centos/7/bs-dmcrypt-journal-colloc/Vagrantfile [new symlink]
tests/functional/centos/7/bs-dmcrypt-journal-colloc/group_vars/all [new file with mode: 0644]
tests/functional/centos/7/bs-dmcrypt-journal-colloc/group_vars/osds [new file with mode: 0644]
tests/functional/centos/7/bs-dmcrypt-journal-colloc/hosts [new file with mode: 0644]
tests/functional/centos/7/bs-dmcrypt-journal-colloc/vagrant_variables.yml [new file with mode: 0644]
tests/functional/centos/7/bs-docker-cluster-dedicated-journal/Vagrantfile [new symlink]
tests/functional/centos/7/bs-docker-cluster-dedicated-journal/group_vars/all [new file with mode: 0644]
tests/functional/centos/7/bs-docker-cluster-dedicated-journal/hosts [new file with mode: 0644]
tests/functional/centos/7/bs-docker-cluster-dedicated-journal/vagrant_variables.yml [new file with mode: 0644]
tests/functional/centos/7/bs-docker-cluster/Vagrantfile [new symlink]
tests/functional/centos/7/bs-docker-cluster/group_vars/all [new file with mode: 0644]
tests/functional/centos/7/bs-docker-cluster/hosts [new file with mode: 0644]
tests/functional/centos/7/bs-docker-cluster/vagrant_variables.yml [new file with mode: 0644]
tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/Vagrantfile [new symlink]
tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/group_vars/all [new file with mode: 0644]
tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/group_vars/osds [new file with mode: 0644]
tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/hosts [new file with mode: 0644]
tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/vagrant_variables.yml [new file with mode: 0644]
tests/functional/centos/7/bs-journal-collocation/vagrant_variables.yml
tests/functional/centos/7/cluster/group_vars/all
tests/functional/centos/7/dmcrypt-dedicated-journal/group_vars/all
tests/functional/centos/7/dmcrypt-journal-collocation/group_vars/all
tests/functional/centos/7/docker-cluster-dedicated-journal/Vagrantfile [changed from file to symlink]
tests/functional/centos/7/docker-cluster-dmcrypt-journal-collocation/Vagrantfile [changed from file to symlink]
tests/functional/centos/7/docker-cluster/group_vars/all
tests/functional/centos/7/journal-collocation/group_vars/all
tox.ini

index 3fa25cf3b223553806dfda0967d025df912c75a9..9d274eb8c103190357d39663af7edd0cdf9eeef9 100644 (file)
@@ -504,7 +504,7 @@ Vagrant.configure(VAGRANTFILE_API_VERSION) do |config|
         # always make /dev/sd{a/b/c} so that CI can ensure that
         # virtualbox and libvirt will have the same devices to use for OSDs
         (0..2).each do |d|
-          lv.storage :file, :device => "hd#{driverletters[d]}", :path => "disk-#{i}-#{d}-#{DISK_UUID}.disk", :size => '12G', :bus => "ide"
+          lv.storage :file, :device => "hd#{driverletters[d]}", :path => "disk-#{i}-#{d}-#{DISK_UUID}.disk", :size => '50G', :bus => "ide"
         end
         lv.memory = MEMORY
         lv.random_hostname = true
index 40f3c4252a13511679cf4c84c651a0886a217627..5dcacfca96db6d863b9d44154303e3f31d8591c6 100644 (file)
@@ -225,13 +225,13 @@ dummy:
 #
 #
 #ceph_osd_docker_devices: "{{ devices }}"
-#ceph_osd_docker_prepare_env: -e CLUSTER={{ cluster }} -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_FORCE_ZAP=1 -e OSD_BLUESTORE=1
+#ceph_osd_docker_prepare_env: -e CLUSTER={{ cluster }} -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_FORCE_ZAP=1
 
 # ACTIVATE DEVICE
 # Examples:
 # Journal collocated or Dedicated journal on Filesore: ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_FILESTORE=1
 # Encrypted OSD or Encrypted OSD with dedicated journal: ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_DMCRYPT=1 -e OSD_FILESTORE=1
-# Bluestore OSD: ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_BLUESTORE=1
+# Bluestore OSD: ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_BLUESTORE=1
 #
 #ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e OSD_JOURNAL_SIZE={{ journal_size }}
 #ceph_osd_docker_run_script_path: "/usr/share" # script called by systemd to run the docker command
index db2f07c109a989f508774d6db1d224d95afa1e71..5aef8bb5095ee7331d82f8a5d0a9296730cf9e50 100644 (file)
@@ -217,13 +217,13 @@ ceph_config_keys: [] # DON'T TOUCH ME
 #
 #
 ceph_osd_docker_devices: "{{ devices }}"
-ceph_osd_docker_prepare_env: -e CLUSTER={{ cluster }} -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_FORCE_ZAP=1 -e OSD_BLUESTORE=1
+ceph_osd_docker_prepare_env: -e CLUSTER={{ cluster }} -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_FORCE_ZAP=1
 
 # ACTIVATE DEVICE
 # Examples:
 # Journal collocated or Dedicated journal on Filesore: ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_FILESTORE=1
 # Encrypted OSD or Encrypted OSD with dedicated journal: ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_DMCRYPT=1 -e OSD_FILESTORE=1
-# Bluestore OSD: ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_BLUESTORE=1
+# Bluestore OSD: ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_BLUESTORE=1
 #
 ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e OSD_JOURNAL_SIZE={{ journal_size }}
 ceph_osd_docker_run_script_path: "/usr/share" # script called by systemd to run the docker command
index b78ae73be95ce4eeb5db5c5c851b7fde4239f456..ae7fb11df945312c25a367a28ad923a2115ef32d 100644 (file)
@@ -9,7 +9,7 @@
 
 # NOTE (alahouze): if the device is a partition, the parted command below has
 # failed, this is why we check if the device is a partition too.
-- name: prepare dmcrypt osd disk(s) with a dedicated journal device
+- name: prepare dmcrypt osd disk(s) with a dedicated journal device (filestore)
   command: "ceph-disk prepare --dmcrypt --cluster {{ cluster }} {{ item.2 }} {{ item.3 }}"
   with_together:
     - "{{ parted_results.results }}"
@@ -26,7 +26,7 @@
     - osd_objectstore == 'filestore'
     - dmcrypt_dedicated_journal
 
-- name: prepare dmcrypt osd disk(s) with a dedicated journal device
+- name: prepare dmcrypt osd disk(s) with a dedicated journal device (bluestore)
   command: "ceph-disk prepare --bluestore --dmcrypt --cluster {{ cluster }} {{ item.2 }} {{ item.3 }}"
   with_together:
     - "{{ parted_results.results }}"
diff --git a/tests/functional/centos/7/bs-dmcrypt-dedicated-journal/Vagrantfile b/tests/functional/centos/7/bs-dmcrypt-dedicated-journal/Vagrantfile
new file mode 120000 (symlink)
index 0000000..dfd7436
--- /dev/null
@@ -0,0 +1 @@
+../../../../../Vagrantfile
\ No newline at end of file
diff --git a/tests/functional/centos/7/bs-dmcrypt-dedicated-journal/group_vars/all b/tests/functional/centos/7/bs-dmcrypt-dedicated-journal/group_vars/all
new file mode 100644 (file)
index 0000000..21ac1f7
--- /dev/null
@@ -0,0 +1,23 @@
+---
+
+ceph_stable: True
+public_network: "192.168.11.0/24"
+cluster_network: "192.168.12.0/24"
+journal_size: 100
+monitor_interface: eth1
+dmcrypt_dedicated_journal: true
+osd_objectstore: "bluestore"
+devices:
+  - '/dev/sda'
+raw_journal_devices:
+  - '/dev/sdb'
+os_tuning_params:
+  - { name: kernel.pid_max, value: 4194303 }
+  - { name: fs.file-max, value: 26234859 }
+ceph_conf_overrides:
+  global:
+    osd_pool_default_pg_num: 8
+    osd_pool_default_size: 1
+  osd:
+    bluestore block db size = 67108864
+    bluestore block wal size = 1048576000
diff --git a/tests/functional/centos/7/bs-dmcrypt-dedicated-journal/group_vars/osds b/tests/functional/centos/7/bs-dmcrypt-dedicated-journal/group_vars/osds
new file mode 100644 (file)
index 0000000..ed9b23a
--- /dev/null
@@ -0,0 +1,4 @@
+---
+os_tuning_params:
+  - { name: kernel.pid_max, value: 4194303 }
+  - { name: fs.file-max, value: 26234859 }
diff --git a/tests/functional/centos/7/bs-dmcrypt-dedicated-journal/hosts b/tests/functional/centos/7/bs-dmcrypt-dedicated-journal/hosts
new file mode 100644 (file)
index 0000000..f6a265a
--- /dev/null
@@ -0,0 +1,5 @@
+[mons]
+mon0
+
+[osds]
+osd0
diff --git a/tests/functional/centos/7/bs-dmcrypt-dedicated-journal/vagrant_variables.yml b/tests/functional/centos/7/bs-dmcrypt-dedicated-journal/vagrant_variables.yml
new file mode 100644 (file)
index 0000000..8a8fb89
--- /dev/null
@@ -0,0 +1,74 @@
+---
+
+# DEPLOY CONTAINERIZED DAEMONS
+docker: false
+
+# DEFINE THE NUMBER OF VMS TO RUN
+mon_vms: 1
+osd_vms: 1
+mds_vms: 0
+rgw_vms: 0
+nfs_vms: 0
+rbd_mirror_vms: 0
+client_vms: 0
+iscsi_gw_vms: 0
+mgr_vms: 0
+
+# Deploy RESTAPI on each of the Monitors
+restapi: true
+
+# INSTALL SOURCE OF CEPH
+# valid values are 'stable' and 'dev'
+ceph_install_source: stable
+
+# SUBNETS TO USE FOR THE VMS
+public_subnet: 192.168.11
+cluster_subnet: 192.168.12
+
+# MEMORY
+# set 1024 for CentOS
+memory: 512
+
+# Ethernet interface name
+# use eth1 for libvirt and ubuntu precise, enp0s8 for CentOS and ubuntu xenial
+eth: 'eth1'
+
+# Disks
+# For libvirt use disks: "[ '/dev/vdb', '/dev/vdc' ]"
+# For CentOS7 use disks: "[ '/dev/sda', '/dev/sdb' ]"
+disks: "[ '/dev/sdb', '/dev/sdc' ]"
+
+# VAGRANT BOX
+# Ceph boxes are *strongly* suggested. They are under better control and will
+# not get updated frequently unless required for build systems. These are (for
+# now):
+#
+# * ceph/ubuntu-xenial
+#
+# Ubuntu: ceph/ubuntu-xenial bento/ubuntu-16.04 or ubuntu/trusty64 or ubuntu/wily64
+# CentOS: bento/centos-7.1 or puppetlabs/centos-7.0-64-puppet
+# libvirt CentOS: centos/7
+# parallels Ubuntu: parallels/ubuntu-14.04
+# Debian: deb/jessie-amd64 - be careful the storage controller is named 'SATA Controller'
+# For more boxes have a look at:
+#   - https://atlas.hashicorp.com/boxes/search?utf8=✓&sort=&provider=virtualbox&q=
+#   - https://download.gluster.org/pub/gluster/purpleidea/vagrant/
+vagrant_box: centos/7
+#ssh_private_key_path: "~/.ssh/id_rsa"
+# The sync directory changes based on vagrant box
+# Set to /home/vagrant/sync for Centos/7, /home/{ user }/vagrant for openstack and defaults to /vagrant
+#vagrant_sync_dir: /home/vagrant/sync
+#vagrant_sync_dir: /
+# Disables synced folder creation. Not needed for testing, will skip mounting
+# the vagrant directory on the remote box regardless of the provider.
+vagrant_disable_synced_folder: true
+# VAGRANT URL
+# This is a URL to download an image from an alternate location.  vagrant_box
+# above should be set to the filename of the image.
+# Fedora virtualbox: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-virtualbox.box
+# Fedora libvirt: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-libvirt.box
+# vagrant_box_url: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-virtualbox.box
+
+os_tuning_params:
+  - { name: kernel.pid_max, value: 4194303 }
+  - { name: fs.file-max, value: 26234859 }
diff --git a/tests/functional/centos/7/bs-dmcrypt-journal-colloc/Vagrantfile b/tests/functional/centos/7/bs-dmcrypt-journal-colloc/Vagrantfile
new file mode 120000 (symlink)
index 0000000..dfd7436
--- /dev/null
@@ -0,0 +1 @@
+../../../../../Vagrantfile
\ No newline at end of file
diff --git a/tests/functional/centos/7/bs-dmcrypt-journal-colloc/group_vars/all b/tests/functional/centos/7/bs-dmcrypt-journal-colloc/group_vars/all
new file mode 100644 (file)
index 0000000..4371089
--- /dev/null
@@ -0,0 +1,22 @@
+---
+
+ceph_stable: True
+public_network: "192.168.13.0/24"
+cluster_network: "192.168.14.0/24"
+journal_size: 100
+monitor_interface: eth1
+dmcrypt_journal_collocation: true
+osd_objectstore: "bluestore"
+devices:
+  - '/dev/sda'
+  - '/dev/sdb'
+os_tuning_params:
+  - { name: kernel.pid_max, value: 4194303 }
+  - { name: fs.file-max, value: 26234859 }
+ceph_conf_overrides:
+  global:
+    osd_pool_default_pg_num: 8
+    osd_pool_default_size: 1
+  osd:
+    bluestore block db size = 67108864
+    bluestore block wal size = 1048576000
diff --git a/tests/functional/centos/7/bs-dmcrypt-journal-colloc/group_vars/osds b/tests/functional/centos/7/bs-dmcrypt-journal-colloc/group_vars/osds
new file mode 100644 (file)
index 0000000..ed9b23a
--- /dev/null
@@ -0,0 +1,4 @@
+---
+os_tuning_params:
+  - { name: kernel.pid_max, value: 4194303 }
+  - { name: fs.file-max, value: 26234859 }
diff --git a/tests/functional/centos/7/bs-dmcrypt-journal-colloc/hosts b/tests/functional/centos/7/bs-dmcrypt-journal-colloc/hosts
new file mode 100644 (file)
index 0000000..f6a265a
--- /dev/null
@@ -0,0 +1,5 @@
+[mons]
+mon0
+
+[osds]
+osd0
diff --git a/tests/functional/centos/7/bs-dmcrypt-journal-colloc/vagrant_variables.yml b/tests/functional/centos/7/bs-dmcrypt-journal-colloc/vagrant_variables.yml
new file mode 100644 (file)
index 0000000..33ff53f
--- /dev/null
@@ -0,0 +1,74 @@
+---
+
+# DEPLOY CONTAINERIZED DAEMONS
+docker: false
+
+# DEFINE THE NUMBER OF VMS TO RUN
+mon_vms: 1
+osd_vms: 1
+mds_vms: 0
+rgw_vms: 0
+nfs_vms: 0
+rbd_mirror_vms: 0
+client_vms: 0
+iscsi_gw_vms: 0
+mgr_vms: 0
+
+# Deploy RESTAPI on each of the Monitors
+restapi: true
+
+# INSTALL SOURCE OF CEPH
+# valid values are 'stable' and 'dev'
+ceph_install_source: stable
+
+# SUBNETS TO USE FOR THE VMS
+public_subnet: 192.168.13
+cluster_subnet: 192.168.14
+
+# MEMORY
+# set 1024 for CentOS
+memory: 512
+
+# Ethernet interface name
+# use eth1 for libvirt and ubuntu precise, enp0s8 for CentOS and ubuntu xenial
+eth: 'eth1'
+
+# Disks
+# For libvirt use disks: "[ '/dev/vdb', '/dev/vdc' ]"
+# For CentOS7 use disks: "[ '/dev/sda', '/dev/sdb' ]"
+disks: "[ '/dev/sdb', '/dev/sdc' ]"
+
+# VAGRANT BOX
+# Ceph boxes are *strongly* suggested. They are under better control and will
+# not get updated frequently unless required for build systems. These are (for
+# now):
+#
+# * ceph/ubuntu-xenial
+#
+# Ubuntu: ceph/ubuntu-xenial bento/ubuntu-16.04 or ubuntu/trusty64 or ubuntu/wily64
+# CentOS: bento/centos-7.1 or puppetlabs/centos-7.0-64-puppet
+# libvirt CentOS: centos/7
+# parallels Ubuntu: parallels/ubuntu-14.04
+# Debian: deb/jessie-amd64 - be careful the storage controller is named 'SATA Controller'
+# For more boxes have a look at:
+#   - https://atlas.hashicorp.com/boxes/search?utf8=✓&sort=&provider=virtualbox&q=
+#   - https://download.gluster.org/pub/gluster/purpleidea/vagrant/
+vagrant_box: centos/7
+#ssh_private_key_path: "~/.ssh/id_rsa"
+# The sync directory changes based on vagrant box
+# Set to /home/vagrant/sync for Centos/7, /home/{ user }/vagrant for openstack and defaults to /vagrant
+#vagrant_sync_dir: /home/vagrant/sync
+#vagrant_sync_dir: /
+# Disables synced folder creation. Not needed for testing, will skip mounting
+# the vagrant directory on the remote box regardless of the provider.
+vagrant_disable_synced_folder: true
+# VAGRANT URL
+# This is a URL to download an image from an alternate location.  vagrant_box
+# above should be set to the filename of the image.
+# Fedora virtualbox: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-virtualbox.box
+# Fedora libvirt: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-libvirt.box
+# vagrant_box_url: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-virtualbox.box
+
+os_tuning_params:
+  - { name: kernel.pid_max, value: 4194303 }
+  - { name: fs.file-max, value: 26234859 }
diff --git a/tests/functional/centos/7/bs-docker-cluster-dedicated-journal/Vagrantfile b/tests/functional/centos/7/bs-docker-cluster-dedicated-journal/Vagrantfile
new file mode 120000 (symlink)
index 0000000..dfd7436
--- /dev/null
@@ -0,0 +1 @@
+../../../../../Vagrantfile
\ No newline at end of file
diff --git a/tests/functional/centos/7/bs-docker-cluster-dedicated-journal/group_vars/all b/tests/functional/centos/7/bs-docker-cluster-dedicated-journal/group_vars/all
new file mode 100644 (file)
index 0000000..09176dc
--- /dev/null
@@ -0,0 +1,23 @@
+---
+# this is only here to let the CI tests know
+# that this scenario is using docker
+docker: True
+
+ceph_stable: True
+containerized_deployment: True
+cluster: test
+monitor_interface: eth1
+ceph_mon_docker_subnet: "{{ public_network }}"
+journal_size: 100
+ceph_docker_on_openstack: False
+public_network: "192.168.15.0/24"
+cluster_network: "192.168.16.0/24"
+ceph_rgw_civetweb_port: 8080
+ceph_osd_docker_devices: "{{ devices }}"
+devices:
+  - /dev/sda
+raw_journal_devices:
+  - /dev/sdb
+ceph_osd_docker_prepare_env: -e CLUSTER={{ cluster }} -e OSD_FORCE_ZAP=1 -e OSD_JOURNAL={{ raw_journal_devices[0] }} -e OSD_BLUESTORE=1
+ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_BLUESTORE=1
+ceph_osd_docker_run_script_path: /var/tmp
diff --git a/tests/functional/centos/7/bs-docker-cluster-dedicated-journal/hosts b/tests/functional/centos/7/bs-docker-cluster-dedicated-journal/hosts
new file mode 100644 (file)
index 0000000..f6a265a
--- /dev/null
@@ -0,0 +1,5 @@
+[mons]
+mon0
+
+[osds]
+osd0
diff --git a/tests/functional/centos/7/bs-docker-cluster-dedicated-journal/vagrant_variables.yml b/tests/functional/centos/7/bs-docker-cluster-dedicated-journal/vagrant_variables.yml
new file mode 100644 (file)
index 0000000..aa03b19
--- /dev/null
@@ -0,0 +1,52 @@
+---
+
+# DEPLOY CONTAINERIZED DAEMONS
+docker: True
+
+# DEFINE THE NUMBER OF VMS TO RUN
+mon_vms: 1
+osd_vms: 1
+mds_vms: 0
+rgw_vms: 0
+nfs_vms: 0
+rbd_mirror_vms: 0
+client_vms: 0
+iscsi_gw_vms: 0
+mgr_vms: 0
+
+# Deploy RESTAPI on each of the Monitors
+restapi: true
+
+# SUBNETS TO USE FOR THE VMS
+public_subnet: 192.168.15
+cluster_subnet: 192.168.16
+
+# MEMORY
+# set 1024 for CentOS
+memory: 1024
+
+# Disks
+# For libvirt use disks: "[ '/dev/vdb', '/dev/vdc' ]"
+# For CentOS7 use disks: "[ '/dev/sda', '/dev/sdb' ]"
+disks: "[ '/dev/sda', '/dev/sdb' ]"
+
+# VAGRANT BOX
+# Ceph boxes are *strongly* suggested. They are under better control and will
+# not get updated frequently unless required for build systems. These are (for
+# now):
+#
+# * ceph/ubuntu-xenial
+#
+# NOTE(leseb): we use centos for this scenario since we at least need Docker version 1.12.5
+# which is not available in Atomic Host.
+# There are bug like this one: https://github.com/docker/docker/issues/12694
+vagrant_box: centos/7
+
+#ssh_private_key_path: "~/.ssh/id_rsa"
+# The sync directory changes based on vagrant box
+# Set to /home/vagrant/sync for Centos/7, /home/{ user }/vagrant for openstack and defaults to /vagrant
+#vagrant_sync_dir: /home/vagrant/sync
+#vagrant_sync_dir: /
+# Disables synced folder creation. Not needed for testing, will skip mounting
+# the vagrant directory on the remote box regardless of the provider.
+vagrant_disable_synced_folder: true
diff --git a/tests/functional/centos/7/bs-docker-cluster/Vagrantfile b/tests/functional/centos/7/bs-docker-cluster/Vagrantfile
new file mode 120000 (symlink)
index 0000000..dfd7436
--- /dev/null
@@ -0,0 +1 @@
+../../../../../Vagrantfile
\ No newline at end of file
diff --git a/tests/functional/centos/7/bs-docker-cluster/group_vars/all b/tests/functional/centos/7/bs-docker-cluster/group_vars/all
new file mode 100644 (file)
index 0000000..bc10cd1
--- /dev/null
@@ -0,0 +1,29 @@
+---
+# this is only here to let the CI tests know
+# that this scenario is using docker
+docker: True
+
+ceph_stable: True
+containerized_deployment: True
+cluster: test
+monitor_interface: eth1
+ceph_mon_docker_subnet: "{{ public_network }}"
+journal_size: 100
+ceph_docker_on_openstack: False
+public_network: "192.168.15.0/24"
+cluster_network: "192.168.16.0/24"
+journal_collocation: true
+ceph_rgw_civetweb_port: 8080
+ceph_osd_docker_devices: "{{ devices }}"
+ceph_osd_docker_prepare_env: -e CLUSTER={{ cluster }} -e OSD_FORCE_ZAP=1 -e OSD_BLUESTORE=1
+ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_BLUESTORE=1
+devices:
+  - /dev/sda
+  - /dev/sdb
+ceph_osd_docker_run_script_path: /var/tmp
+rgw_override_bucket_index_max_shards: 16
+rgw_bucket_default_quota_max_objects: 1638400
+ceph_conf_overrides:
+  global:
+    osd_pool_default_pg_num: 8
+    osd_pool_default_size: 1
diff --git a/tests/functional/centos/7/bs-docker-cluster/hosts b/tests/functional/centos/7/bs-docker-cluster/hosts
new file mode 100644 (file)
index 0000000..2d68436
--- /dev/null
@@ -0,0 +1,16 @@
+[mons]
+mon0
+mon1
+mon2
+
+[osds]
+osd0
+
+[mdss]
+mds0
+
+[rgws]
+rgw0
+
+[mgrs]
+mgr0
diff --git a/tests/functional/centos/7/bs-docker-cluster/vagrant_variables.yml b/tests/functional/centos/7/bs-docker-cluster/vagrant_variables.yml
new file mode 100644 (file)
index 0000000..5c4d108
--- /dev/null
@@ -0,0 +1,62 @@
+---
+
+# DEPLOY CONTAINERIZED DAEMONS
+docker: True
+
+# DEFINE THE NUMBER OF VMS TO RUN
+mon_vms: 3
+osd_vms: 1
+mds_vms: 1
+rgw_vms: 1
+nfs_vms: 0
+rbd_mirror_vms: 1
+client_vms: 0
+iscsi_gw_vms: 0
+mgr_vms: 1
+
+# Deploy RESTAPI on each of the Monitors
+restapi: true
+
+# SUBNETS TO USE FOR THE VMS
+public_subnet: 192.168.15
+cluster_subnet: 192.168.16
+
+# MEMORY
+# set 1024 for CentOS
+memory: 1024
+
+# Disks
+# For libvirt use disks: "[ '/dev/vdb', '/dev/vdc' ]"
+# For CentOS7 use disks: "[ '/dev/sda', '/dev/sdb' ]"
+disks: "[ '/dev/sda', '/dev/sdb' ]"
+
+# VAGRANT BOX
+# Ceph boxes are *strongly* suggested. They are under better control and will
+# not get updated frequently unless required for build systems. These are (for
+# now):
+#
+# * ceph/ubuntu-xenial
+#
+# Ubuntu: ceph/ubuntu-xenial bento/ubuntu-16.04 or ubuntu/trusty64 or ubuntu/wily64
+# CentOS: bento/centos-7.1 or puppetlabs/centos-7.0-64-puppet
+# libvirt CentOS: centos/7
+# parallels Ubuntu: parallels/ubuntu-14.04
+# Debian: deb/jessie-amd64 - be careful the storage controller is named 'SATA Controller'
+# For more boxes have a look at:
+#   - https://atlas.hashicorp.com/boxes/search?utf8=✓&sort=&provider=virtualbox&q=
+#   - https://download.gluster.org/pub/gluster/purpleidea/vagrant/
+vagrant_box: centos/atomic-host
+#ssh_private_key_path: "~/.ssh/id_rsa"
+# The sync directory changes based on vagrant box
+# Set to /home/vagrant/sync for Centos/7, /home/{ user }/vagrant for openstack and defaults to /vagrant
+#vagrant_sync_dir: /home/vagrant/sync
+#vagrant_sync_dir: /
+# Disables synced folder creation. Not needed for testing, will skip mounting
+# the vagrant directory on the remote box regardless of the provider.
+vagrant_disable_synced_folder: true
+# VAGRANT URL
+# This is a URL to download an image from an alternate location.  vagrant_box
+# above should be set to the filename of the image.
+# Fedora virtualbox: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-virtualbox.box
+# Fedora libvirt: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-libvirt.box
+# vagrant_box_url: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-virtualbox.box
diff --git a/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/Vagrantfile b/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/Vagrantfile
new file mode 120000 (symlink)
index 0000000..dfd7436
--- /dev/null
@@ -0,0 +1 @@
+../../../../../Vagrantfile
\ No newline at end of file
diff --git a/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/group_vars/all b/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/group_vars/all
new file mode 100644 (file)
index 0000000..6a2a585
--- /dev/null
@@ -0,0 +1,26 @@
+---
+docker: true
+containerized_deployment: True
+ceph_stable: True
+ceph_mon_docker_subnet: "{{ public_network }}"
+public_network: "192.168.23.0/24"
+cluster_network: "192.168.24.0/24"
+journal_size: 100
+monitor_interface: eth1
+dmcrypt_journal_collocation: true
+osd_objectstore: "bluestore"
+devices:
+  - '/dev/sda'
+  - '/dev/sdb'
+os_tuning_params:
+  - { name: kernel.pid_max, value: 4194303 }
+  - { name: fs.file-max, value: 26234859 }
+ceph_conf_overrides:
+  global:
+    osd_pool_default_pg_num: 8
+    osd_pool_default_size: 1
+  osd:
+    bluestore block db size = 67108864
+    bluestore block wal size = 1048576000
+ceph_osd_docker_prepare_env: -e CLUSTER={{ cluster }} -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_FORCE_ZAP=1 -e OSD_DMCRYPT=1
+ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_DMCRYPT=1
diff --git a/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/group_vars/osds b/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/group_vars/osds
new file mode 100644 (file)
index 0000000..ed9b23a
--- /dev/null
@@ -0,0 +1,4 @@
+---
+os_tuning_params:
+  - { name: kernel.pid_max, value: 4194303 }
+  - { name: fs.file-max, value: 26234859 }
diff --git a/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/hosts b/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/hosts
new file mode 100644 (file)
index 0000000..f6a265a
--- /dev/null
@@ -0,0 +1,5 @@
+[mons]
+mon0
+
+[osds]
+osd0
diff --git a/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/vagrant_variables.yml b/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc/vagrant_variables.yml
new file mode 100644 (file)
index 0000000..6e03fab
--- /dev/null
@@ -0,0 +1,74 @@
+---
+
+# DEPLOY CONTAINERIZED DAEMONS
+docker: true
+
+# DEFINE THE NUMBER OF VMS TO RUN
+mon_vms: 1
+osd_vms: 1
+mds_vms: 0
+rgw_vms: 0
+nfs_vms: 0
+rbd_mirror_vms: 0
+client_vms: 0
+iscsi_gw_vms: 0
+mgr_vms: 0
+
+# Deploy RESTAPI on each of the Monitors
+restapi: true
+
+# INSTALL SOURCE OF CEPH
+# valid values are 'stable' and 'dev'
+ceph_install_source: stable
+
+# SUBNETS TO USE FOR THE VMS
+public_subnet: 192.168.23
+cluster_subnet: 192.168.24
+
+# MEMORY
+# set 1024 for CentOS
+memory: 512
+
+# Ethernet interface name
+# use eth1 for libvirt and ubuntu precise, enp0s8 for CentOS and ubuntu xenial
+eth: 'eth1'
+
+# Disks
+# For libvirt use disks: "[ '/dev/vdb', '/dev/vdc' ]"
+# For CentOS7 use disks: "[ '/dev/sda', '/dev/sdb' ]"
+disks: "[ '/dev/sdb', '/dev/sdc' ]"
+
+# VAGRANT BOX
+# Ceph boxes are *strongly* suggested. They are under better control and will
+# not get updated frequently unless required for build systems. These are (for
+# now):
+#
+# * ceph/ubuntu-xenial
+#
+# Ubuntu: ceph/ubuntu-xenial bento/ubuntu-16.04 or ubuntu/trusty64 or ubuntu/wily64
+# CentOS: bento/centos-7.1 or puppetlabs/centos-7.0-64-puppet
+# libvirt CentOS: centos/7
+# parallels Ubuntu: parallels/ubuntu-14.04
+# Debian: deb/jessie-amd64 - be careful the storage controller is named 'SATA Controller'
+# For more boxes have a look at:
+#   - https://atlas.hashicorp.com/boxes/search?utf8=✓&sort=&provider=virtualbox&q=
+#   - https://download.gluster.org/pub/gluster/purpleidea/vagrant/
+vagrant_box: centos/7
+#ssh_private_key_path: "~/.ssh/id_rsa"
+# The sync directory changes based on vagrant box
+# Set to /home/vagrant/sync for Centos/7, /home/{ user }/vagrant for openstack and defaults to /vagrant
+#vagrant_sync_dir: /home/vagrant/sync
+#vagrant_sync_dir: /
+# Disables synced folder creation. Not needed for testing, will skip mounting
+# the vagrant directory on the remote box regardless of the provider.
+vagrant_disable_synced_folder: true
+# VAGRANT URL
+# This is a URL to download an image from an alternate location.  vagrant_box
+# above should be set to the filename of the image.
+# Fedora virtualbox: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-virtualbox.box
+# Fedora libvirt: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-libvirt.box
+# vagrant_box_url: https://download.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-virtualbox.box
+
+os_tuning_params:
+  - { name: kernel.pid_max, value: 4194303 }
+  - { name: fs.file-max, value: 26234859 }
index cbb34384216ba0456e78e573cc31bfad423ccdc2..625cbe8caeacc49ef187047ae2e01721115b6cde 100644 (file)
@@ -38,6 +38,9 @@ eth: 'eth1'
 # For CentOS7 use disks: "[ '/dev/sda', '/dev/sdb' ]"
 disks: "[ '/dev/sdb', '/dev/sdc' ]"
 
+# VM disk size in MB
+disk_size: 41000
+
 # VAGRANT BOX
 # Ceph boxes are *strongly* suggested. They are under better control and will
 # not get updated frequently unless required for build systems. These are (for
index d13902a7773afed7a9e903e6c66a0cfd8ba39f94..9f7313efd77442779b8ba3028513a400784120ee 100644 (file)
@@ -6,6 +6,7 @@ public_network: "192.168.1.0/24"
 cluster_network: "192.168.2.0/24"
 monitor_interface: eth1
 journal_size: 100
+osd_objectstore: "filestore"
 devices:
   - '/dev/sda'
 raw_journal_devices:
index f5356b161e9a5a0960eb43fa2d264ddfe69f4008..6313dcac6b36ba2b56e103744dde96c6a36084ef 100644 (file)
@@ -6,6 +6,7 @@ cluster_network: "192.168.12.0/24"
 journal_size: 100
 monitor_interface: eth1
 dmcrypt_dedicated_journal: true
+osd_objectstore: "filestore"
 devices:
   - '/dev/sda'
 raw_journal_devices:
index 0795478a93a2c15ec40e381492cfbcc59cef0b9a..a4d740c3990aeb0b966306f1f0d5ef963e1d0076 100644 (file)
@@ -6,6 +6,7 @@ cluster_network: "192.168.14.0/24"
 journal_size: 100
 monitor_interface: eth1
 dmcrypt_journal_collocation: true
+osd_objectstore: "filestore"
 devices:
   - '/dev/sda'
   - '/dev/sdb'
deleted file mode 100644 (file)
index 0ba579791098950c468efcbd97b54fe9651150e0..0000000000000000000000000000000000000000
+++ /dev/null
@@ -1,492 +0,0 @@
-# -*- mode: ruby -*-
-# vi: set ft=ruby :
-
-require 'yaml'
-require 'time'
-VAGRANTFILE_API_VERSION = '2'
-
-DEBUG = false
-
-config_file=File.expand_path(File.join(File.dirname(__FILE__), 'vagrant_variables.yml'))
-settings=YAML.load_file(config_file)
-
-LABEL_PREFIX   = settings['label_prefix'] ? settings['label_prefix'] + "-" : ""
-NMONS          = settings['mon_vms']
-NOSDS          = settings['osd_vms']
-NMDSS          = settings['mds_vms']
-NRGWS          = settings['rgw_vms']
-NNFSS          = settings['nfs_vms']
-RESTAPI        = settings['restapi']
-NRBD_MIRRORS   = settings['rbd_mirror_vms']
-CLIENTS        = settings['client_vms']
-NISCSI_GWS     = settings['iscsi_gw_vms']
-PUBLIC_SUBNET  = settings['public_subnet']
-CLUSTER_SUBNET = settings['cluster_subnet']
-BOX            = settings['vagrant_box']
-BOX_URL        = settings['vagrant_box_url']
-SYNC_DIR       = settings['vagrant_sync_dir']
-MEMORY         = settings['memory']
-ETH            = settings['eth']
-DOCKER         = settings['docker']
-USER           = settings['ssh_username']
-
-ASSIGN_STATIC_IP = !(BOX == 'openstack' or BOX == 'linode')
-DISABLE_SYNCED_FOLDER = settings.fetch('vagrant_disable_synced_folder', false)
-DISK_UUID = Time.now.utc.to_i
-
-
-ansible_provision = proc do |ansible|
-  if DOCKER then
-    ansible.playbook = 'site-docker.yml'
-    if settings['skip_tags']
-      ansible.skip_tags = settings['skip_tags']
-    end
-  else
-    ansible.playbook = 'site.yml'
-  end
-
-  # Note: Can't do ranges like mon[0-2] in groups because
-  # these aren't supported by Vagrant, see
-  # https://github.com/mitchellh/vagrant/issues/3539
-  ansible.groups = {
-    'mons'            => (0..NMONS - 1).map { |j| "#{LABEL_PREFIX}mon#{j}" },
-    'osds'            => (0..NOSDS - 1).map { |j| "#{LABEL_PREFIX}osd#{j}" },
-    'mdss'            => (0..NMDSS - 1).map { |j| "#{LABEL_PREFIX}mds#{j}" },
-    'rgws'            => (0..NRGWS - 1).map { |j| "#{LABEL_PREFIX}rgw#{j}" },
-    'nfss'            => (0..NNFSS - 1).map { |j| "#{LABEL_PREFIX}nfs#{j}" },
-    'rbd_mirrors'     => (0..NRBD_MIRRORS - 1).map { |j| "#{LABEL_PREFIX}rbd_mirror#{j}" },
-    'clients'         => (0..CLIENTS - 1).map { |j| "#{LABEL_PREFIX}client#{j}" },
-    'iscsi_gw'        => (0..NISCSI_GWS - 1).map { |j| "#{LABEL_PREFIX}iscsi_gw#{j}" }
-  }
-
-  if RESTAPI then
-    ansible.groups['restapis'] = (0..NMONS - 1).map { |j| "#{LABEL_PREFIX}mon#{j}" }
-  end
-
-  ansible.extra_vars = {
-      cluster_network: "#{CLUSTER_SUBNET}.0/24",
-      journal_size: 100,
-      public_network: "#{PUBLIC_SUBNET}.0/24",
-  }
-
-  # In a production deployment, these should be secret
-  if DOCKER then
-    ansible.extra_vars = ansible.extra_vars.merge({
-      containerized_deployment: 'true',
-      monitor_interface: ETH,
-      ceph_mon_docker_subnet: "#{PUBLIC_SUBNET}.0/24",
-      ceph_osd_docker_devices: settings['disks'],
-      devices: settings['disks'],
-      ceph_docker_on_openstack: BOX == 'openstack',
-      ceph_rgw_civetweb_port: 8080,
-      generate_fsid: 'true',
-    })
-  else
-    ansible.extra_vars = ansible.extra_vars.merge({
-      devices: settings['disks'],
-      journal_collocation: 'true',
-      monitor_interface: ETH,
-      os_tuning_params: settings['os_tuning_params'],
-      pool_default_size: '2',
-    })
-  end
-
-  if BOX == 'linode' then
-    ansible.sudo = true
-    # Use monitor_address_block instead of monitor_interface:
-    ansible.extra_vars.delete(:monitor_interface)
-    ansible.extra_vars = ansible.extra_vars.merge({
-      cluster_network: "#{CLUSTER_SUBNET}.0/16",
-      devices: ['/dev/sdc'], # hardcode leftover disk
-      journal_collocation: 'true',
-      monitor_address_block: "#{PUBLIC_SUBNET}.0/16",
-      public_network: "#{PUBLIC_SUBNET}.0/16",
-    })
-  end
-
-  if DEBUG then
-    ansible.verbose = '-vvv'
-  end
-  ansible.limit = 'all'
-end
-
-def create_vmdk(name, size)
-  dir = Pathname.new(__FILE__).expand_path.dirname
-  path = File.join(dir, '.vagrant', name + '.vmdk')
-  `vmware-vdiskmanager -c -s #{size} -t 0 -a scsi #{path} \
-   2>&1 > /dev/null` unless File.exist?(path)
-end
-
-Vagrant.configure(VAGRANTFILE_API_VERSION) do |config|
-  config.vm.box = BOX
-  config.vm.box_url = BOX_URL
-  config.ssh.insert_key = false # workaround for https://github.com/mitchellh/vagrant/issues/5048
-  config.ssh.private_key_path = settings['ssh_private_key_path']
-  config.ssh.username = USER
-
-  # Faster bootup. Disables mounting the sync folder for libvirt and virtualbox
-  if DISABLE_SYNCED_FOLDER
-    config.vm.provider :virtualbox do |v,override|
-      override.vm.synced_folder '.', SYNC_DIR, disabled: true
-    end
-    config.vm.provider :libvirt do |v,override|
-      override.vm.synced_folder '.', SYNC_DIR, disabled: true
-    end
-  end
-
-  if BOX == 'openstack'
-    # OpenStack VMs
-    config.vm.provider :openstack do |os|
-      config.vm.synced_folder ".", "/home/#{USER}/vagrant", disabled: true
-      config.ssh.pty = true
-      os.openstack_auth_url = settings['os_openstack_auth_url']
-      os.username = settings['os_username']
-      os.password = settings['os_password']
-      os.tenant_name = settings['os_tenant_name']
-      os.region = settings['os_region']
-      os.flavor = settings['os_flavor']
-      os.image = settings['os_image']
-      os.keypair_name = settings['os_keypair_name']
-      os.security_groups = ['default']
-
-      if settings['os.networks'] then
-        os.networks = settings['os_networks']
-      end
-
-      if settings['os.floating_ip_pool'] then
-        os.floating_ip_pool = settings['os_floating_ip_pool']
-      end
-
-      config.vm.provision "shell", inline: "true", upload_path: "/home/#{USER}/vagrant-shell"
-    end
-  elsif BOX == 'linode'
-    config.vm.provider :linode do |provider, override|
-      provider.token = ENV['LINODE_API_KEY']
-      provider.distribution = settings['cloud_distribution'] # 'Ubuntu 16.04 LTS'
-      provider.datacenter = settings['cloud_datacenter']
-      provider.plan = MEMORY.to_s
-      provider.private_networking = true
-      # root install generally takes <1GB
-      provider.xvda_size = 4*1024
-      # add some swap as the Linode distros require it
-      provider.swap_size = 128
-    end
-  end
-
-  (0..CLIENTS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}client#{i}" do |client|
-      client.vm.hostname = "#{LABEL_PREFIX}ceph-client#{i}"
-      if ASSIGN_STATIC_IP
-        client.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.4#{i}"
-      end
-      # Virtualbox
-      client.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      client.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      client.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-
-      # Parallels
-      client.vm.provider "parallels" do |prl|
-        prl.name = "ceph-client#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      client.vm.provider :linode do |provider|
-        provider.label = client.vm.hostname
-      end
-    end
-  end
-
-  (0..NRGWS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}rgw#{i}" do |rgw|
-      rgw.vm.hostname = "#{LABEL_PREFIX}ceph-rgw#{i}"
-      if ASSIGN_STATIC_IP
-        rgw.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.5#{i}"
-      end
-
-      # Virtualbox
-      rgw.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      rgw.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      rgw.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-
-      # Parallels
-      rgw.vm.provider "parallels" do |prl|
-        prl.name = "ceph-rgw#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      rgw.vm.provider :linode do |provider|
-        provider.label = rgw.vm.hostname
-      end
-    end
-  end
-
-  (0..NNFSS - 1).each do |i|
-    config.vm.define "nfs#{i}" do |nfs|
-      nfs.vm.hostname = "ceph-nfs#{i}"
-      if ASSIGN_STATIC_IP
-        nfs.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.6#{i}"
-      end
-
-      # Virtualbox
-      nfs.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      nfs.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      nfs.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-
-      # Parallels
-      nfs.vm.provider "parallels" do |prl|
-        prl.name = "ceph-nfs#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      nfs.vm.provider :linode do |provider|
-        provider.label = nfs.vm.hostname
-      end
-    end
-  end
-
-  (0..NMDSS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}mds#{i}" do |mds|
-      mds.vm.hostname = "#{LABEL_PREFIX}ceph-mds#{i}"
-      if ASSIGN_STATIC_IP
-        mds.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.7#{i}"
-      end
-      # Virtualbox
-      mds.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      mds.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      mds.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-      # Parallels
-      mds.vm.provider "parallels" do |prl|
-        prl.name = "ceph-mds#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      mds.vm.provider :linode do |provider|
-        provider.label = mds.vm.hostname
-      end
-    end
-  end
-
-  (0..NRBD_MIRRORS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}rbd_mirror#{i}" do |rbd_mirror|
-      rbd_mirror.vm.hostname = "#{LABEL_PREFIX}ceph-rbd-mirror#{i}"
-      if ASSIGN_STATIC_IP
-        rbd_mirror.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.8#{i}"
-      end
-      # Virtualbox
-      rbd_mirror.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      rbd_mirror.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      rbd_mirror.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-      # Parallels
-      rbd_mirror.vm.provider "parallels" do |prl|
-        prl.name = "ceph-rbd-mirror#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      rbd_mirror.vm.provider :linode do |provider|
-        provider.label = rbd_mirror.vm.hostname
-      end
-    end
-  end
-
-  (0..NISCSI_GWS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}iscsi_gw#{i}" do |iscsi_gw|
-      iscsi_gw.vm.hostname = "#{LABEL_PREFIX}ceph-iscsi-gw#{i}"
-      if ASSIGN_STATIC_IP
-        iscsi_gw.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.9#{i}"
-      end
-      # Virtualbox
-      iscsi_gw.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      iscsi_gw.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      iscsi_gw.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-      # Parallels
-      iscsi_gw.vm.provider "parallels" do |prl|
-        prl.name = "ceph-iscsi-gw#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      iscsi_gw.vm.provider :linode do |provider|
-        provider.label = iscsi_gw.vm.hostname
-      end
-    end
-  end
-
-  (0..NMONS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}mon#{i}" do |mon|
-      mon.vm.hostname = "#{LABEL_PREFIX}ceph-mon#{i}"
-      if ASSIGN_STATIC_IP
-        mon.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.1#{i}"
-      end
-      # Virtualbox
-      mon.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      mon.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      mon.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = false
-      end
-
-      # Parallels
-      mon.vm.provider "parallels" do |prl|
-        prl.name = "ceph-mon#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      mon.vm.provider :linode do |provider|
-        provider.label = mon.vm.hostname
-      end
-    end
-  end
-
-  (0..NOSDS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}osd#{i}" do |osd|
-      osd.vm.hostname = "#{LABEL_PREFIX}ceph-osd#{i}"
-      if ASSIGN_STATIC_IP
-        osd.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.10#{i}"
-        osd.vm.network :private_network,
-          ip: "#{CLUSTER_SUBNET}.20#{i}"
-      end
-      # Virtualbox
-      osd.vm.provider :virtualbox do |vb|
-        # Create our own controller for consistency and to remove VM dependency
-        vb.customize ['storagectl', :id,
-                      '--name', 'OSD Controller',
-                      '--add', 'scsi']
-        (0..1).each do |d|
-          vb.customize ['createhd',
-                        '--filename', "disk-#{i}-#{d}",
-                        '--size', '11000'] unless File.exist?("disk-#{i}-#{d}.vdi")
-          vb.customize ['storageattach', :id,
-                        '--storagectl', 'OSD Controller',
-                        '--port', 3 + d,
-                        '--device', 0,
-                        '--type', 'hdd',
-                        '--medium', "disk-#{i}-#{d}.vdi"]
-        end
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      osd.vm.provider :vmware_fusion do |v|
-        (0..1).each do |d|
-          v.vmx["scsi0:#{d + 1}.present"] = 'TRUE'
-          v.vmx["scsi0:#{d + 1}.fileName"] =
-            create_vmdk("disk-#{i}-#{d}", '11000MB')
-        end
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      driverletters = ('a'..'z').to_a
-      osd.vm.provider :libvirt do |lv|
-        # always make /dev/sd{a/b/c} so that CI can ensure that
-        # virtualbox and libvirt will have the same devices to use for OSDs
-        (0..2).each do |d|
-          lv.storage :file, :device => "hd#{driverletters[d]}", :path => "disk-#{i}-#{d}-#{DISK_UUID}.disk", :size => '12G', :bus => "ide"
-        end
-        lv.memory = MEMORY
-        lv.random_hostname = false
-      end
-
-      # Parallels
-      osd.vm.provider "parallels" do |prl|
-        prl.name = "ceph-osd#{i}"
-        prl.memory = "#{MEMORY}"
-        (0..1).each do |d|
-          prl.customize ["set", :id,
-                         "--device-add",
-                         "hdd",
-                         "--iface",
-                         "sata"]
-        end
-      end
-
-      osd.vm.provider :linode do |provider|
-        provider.label = osd.vm.hostname
-      end
-
-      # Run the provisioner after the last machine comes up
-      osd.vm.provision 'ansible', &ansible_provision if i == (NOSDS - 1)
-    end
-  end
-end
new file mode 120000 (symlink)
index 0000000000000000000000000000000000000000..dfd7436c93f56089cc77635c8d49e0ccebcf0d3b
--- /dev/null
@@ -0,0 +1 @@
+../../../../../Vagrantfile
\ No newline at end of file
deleted file mode 100644 (file)
index 0ba579791098950c468efcbd97b54fe9651150e0..0000000000000000000000000000000000000000
+++ /dev/null
@@ -1,492 +0,0 @@
-# -*- mode: ruby -*-
-# vi: set ft=ruby :
-
-require 'yaml'
-require 'time'
-VAGRANTFILE_API_VERSION = '2'
-
-DEBUG = false
-
-config_file=File.expand_path(File.join(File.dirname(__FILE__), 'vagrant_variables.yml'))
-settings=YAML.load_file(config_file)
-
-LABEL_PREFIX   = settings['label_prefix'] ? settings['label_prefix'] + "-" : ""
-NMONS          = settings['mon_vms']
-NOSDS          = settings['osd_vms']
-NMDSS          = settings['mds_vms']
-NRGWS          = settings['rgw_vms']
-NNFSS          = settings['nfs_vms']
-RESTAPI        = settings['restapi']
-NRBD_MIRRORS   = settings['rbd_mirror_vms']
-CLIENTS        = settings['client_vms']
-NISCSI_GWS     = settings['iscsi_gw_vms']
-PUBLIC_SUBNET  = settings['public_subnet']
-CLUSTER_SUBNET = settings['cluster_subnet']
-BOX            = settings['vagrant_box']
-BOX_URL        = settings['vagrant_box_url']
-SYNC_DIR       = settings['vagrant_sync_dir']
-MEMORY         = settings['memory']
-ETH            = settings['eth']
-DOCKER         = settings['docker']
-USER           = settings['ssh_username']
-
-ASSIGN_STATIC_IP = !(BOX == 'openstack' or BOX == 'linode')
-DISABLE_SYNCED_FOLDER = settings.fetch('vagrant_disable_synced_folder', false)
-DISK_UUID = Time.now.utc.to_i
-
-
-ansible_provision = proc do |ansible|
-  if DOCKER then
-    ansible.playbook = 'site-docker.yml'
-    if settings['skip_tags']
-      ansible.skip_tags = settings['skip_tags']
-    end
-  else
-    ansible.playbook = 'site.yml'
-  end
-
-  # Note: Can't do ranges like mon[0-2] in groups because
-  # these aren't supported by Vagrant, see
-  # https://github.com/mitchellh/vagrant/issues/3539
-  ansible.groups = {
-    'mons'            => (0..NMONS - 1).map { |j| "#{LABEL_PREFIX}mon#{j}" },
-    'osds'            => (0..NOSDS - 1).map { |j| "#{LABEL_PREFIX}osd#{j}" },
-    'mdss'            => (0..NMDSS - 1).map { |j| "#{LABEL_PREFIX}mds#{j}" },
-    'rgws'            => (0..NRGWS - 1).map { |j| "#{LABEL_PREFIX}rgw#{j}" },
-    'nfss'            => (0..NNFSS - 1).map { |j| "#{LABEL_PREFIX}nfs#{j}" },
-    'rbd_mirrors'     => (0..NRBD_MIRRORS - 1).map { |j| "#{LABEL_PREFIX}rbd_mirror#{j}" },
-    'clients'         => (0..CLIENTS - 1).map { |j| "#{LABEL_PREFIX}client#{j}" },
-    'iscsi_gw'        => (0..NISCSI_GWS - 1).map { |j| "#{LABEL_PREFIX}iscsi_gw#{j}" }
-  }
-
-  if RESTAPI then
-    ansible.groups['restapis'] = (0..NMONS - 1).map { |j| "#{LABEL_PREFIX}mon#{j}" }
-  end
-
-  ansible.extra_vars = {
-      cluster_network: "#{CLUSTER_SUBNET}.0/24",
-      journal_size: 100,
-      public_network: "#{PUBLIC_SUBNET}.0/24",
-  }
-
-  # In a production deployment, these should be secret
-  if DOCKER then
-    ansible.extra_vars = ansible.extra_vars.merge({
-      containerized_deployment: 'true',
-      monitor_interface: ETH,
-      ceph_mon_docker_subnet: "#{PUBLIC_SUBNET}.0/24",
-      ceph_osd_docker_devices: settings['disks'],
-      devices: settings['disks'],
-      ceph_docker_on_openstack: BOX == 'openstack',
-      ceph_rgw_civetweb_port: 8080,
-      generate_fsid: 'true',
-    })
-  else
-    ansible.extra_vars = ansible.extra_vars.merge({
-      devices: settings['disks'],
-      journal_collocation: 'true',
-      monitor_interface: ETH,
-      os_tuning_params: settings['os_tuning_params'],
-      pool_default_size: '2',
-    })
-  end
-
-  if BOX == 'linode' then
-    ansible.sudo = true
-    # Use monitor_address_block instead of monitor_interface:
-    ansible.extra_vars.delete(:monitor_interface)
-    ansible.extra_vars = ansible.extra_vars.merge({
-      cluster_network: "#{CLUSTER_SUBNET}.0/16",
-      devices: ['/dev/sdc'], # hardcode leftover disk
-      journal_collocation: 'true',
-      monitor_address_block: "#{PUBLIC_SUBNET}.0/16",
-      public_network: "#{PUBLIC_SUBNET}.0/16",
-    })
-  end
-
-  if DEBUG then
-    ansible.verbose = '-vvv'
-  end
-  ansible.limit = 'all'
-end
-
-def create_vmdk(name, size)
-  dir = Pathname.new(__FILE__).expand_path.dirname
-  path = File.join(dir, '.vagrant', name + '.vmdk')
-  `vmware-vdiskmanager -c -s #{size} -t 0 -a scsi #{path} \
-   2>&1 > /dev/null` unless File.exist?(path)
-end
-
-Vagrant.configure(VAGRANTFILE_API_VERSION) do |config|
-  config.vm.box = BOX
-  config.vm.box_url = BOX_URL
-  config.ssh.insert_key = false # workaround for https://github.com/mitchellh/vagrant/issues/5048
-  config.ssh.private_key_path = settings['ssh_private_key_path']
-  config.ssh.username = USER
-
-  # Faster bootup. Disables mounting the sync folder for libvirt and virtualbox
-  if DISABLE_SYNCED_FOLDER
-    config.vm.provider :virtualbox do |v,override|
-      override.vm.synced_folder '.', SYNC_DIR, disabled: true
-    end
-    config.vm.provider :libvirt do |v,override|
-      override.vm.synced_folder '.', SYNC_DIR, disabled: true
-    end
-  end
-
-  if BOX == 'openstack'
-    # OpenStack VMs
-    config.vm.provider :openstack do |os|
-      config.vm.synced_folder ".", "/home/#{USER}/vagrant", disabled: true
-      config.ssh.pty = true
-      os.openstack_auth_url = settings['os_openstack_auth_url']
-      os.username = settings['os_username']
-      os.password = settings['os_password']
-      os.tenant_name = settings['os_tenant_name']
-      os.region = settings['os_region']
-      os.flavor = settings['os_flavor']
-      os.image = settings['os_image']
-      os.keypair_name = settings['os_keypair_name']
-      os.security_groups = ['default']
-
-      if settings['os.networks'] then
-        os.networks = settings['os_networks']
-      end
-
-      if settings['os.floating_ip_pool'] then
-        os.floating_ip_pool = settings['os_floating_ip_pool']
-      end
-
-      config.vm.provision "shell", inline: "true", upload_path: "/home/#{USER}/vagrant-shell"
-    end
-  elsif BOX == 'linode'
-    config.vm.provider :linode do |provider, override|
-      provider.token = ENV['LINODE_API_KEY']
-      provider.distribution = settings['cloud_distribution'] # 'Ubuntu 16.04 LTS'
-      provider.datacenter = settings['cloud_datacenter']
-      provider.plan = MEMORY.to_s
-      provider.private_networking = true
-      # root install generally takes <1GB
-      provider.xvda_size = 4*1024
-      # add some swap as the Linode distros require it
-      provider.swap_size = 128
-    end
-  end
-
-  (0..CLIENTS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}client#{i}" do |client|
-      client.vm.hostname = "#{LABEL_PREFIX}ceph-client#{i}"
-      if ASSIGN_STATIC_IP
-        client.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.4#{i}"
-      end
-      # Virtualbox
-      client.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      client.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      client.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-
-      # Parallels
-      client.vm.provider "parallels" do |prl|
-        prl.name = "ceph-client#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      client.vm.provider :linode do |provider|
-        provider.label = client.vm.hostname
-      end
-    end
-  end
-
-  (0..NRGWS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}rgw#{i}" do |rgw|
-      rgw.vm.hostname = "#{LABEL_PREFIX}ceph-rgw#{i}"
-      if ASSIGN_STATIC_IP
-        rgw.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.5#{i}"
-      end
-
-      # Virtualbox
-      rgw.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      rgw.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      rgw.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-
-      # Parallels
-      rgw.vm.provider "parallels" do |prl|
-        prl.name = "ceph-rgw#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      rgw.vm.provider :linode do |provider|
-        provider.label = rgw.vm.hostname
-      end
-    end
-  end
-
-  (0..NNFSS - 1).each do |i|
-    config.vm.define "nfs#{i}" do |nfs|
-      nfs.vm.hostname = "ceph-nfs#{i}"
-      if ASSIGN_STATIC_IP
-        nfs.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.6#{i}"
-      end
-
-      # Virtualbox
-      nfs.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      nfs.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      nfs.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-
-      # Parallels
-      nfs.vm.provider "parallels" do |prl|
-        prl.name = "ceph-nfs#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      nfs.vm.provider :linode do |provider|
-        provider.label = nfs.vm.hostname
-      end
-    end
-  end
-
-  (0..NMDSS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}mds#{i}" do |mds|
-      mds.vm.hostname = "#{LABEL_PREFIX}ceph-mds#{i}"
-      if ASSIGN_STATIC_IP
-        mds.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.7#{i}"
-      end
-      # Virtualbox
-      mds.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      mds.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      mds.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-      # Parallels
-      mds.vm.provider "parallels" do |prl|
-        prl.name = "ceph-mds#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      mds.vm.provider :linode do |provider|
-        provider.label = mds.vm.hostname
-      end
-    end
-  end
-
-  (0..NRBD_MIRRORS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}rbd_mirror#{i}" do |rbd_mirror|
-      rbd_mirror.vm.hostname = "#{LABEL_PREFIX}ceph-rbd-mirror#{i}"
-      if ASSIGN_STATIC_IP
-        rbd_mirror.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.8#{i}"
-      end
-      # Virtualbox
-      rbd_mirror.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      rbd_mirror.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      rbd_mirror.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-      # Parallels
-      rbd_mirror.vm.provider "parallels" do |prl|
-        prl.name = "ceph-rbd-mirror#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      rbd_mirror.vm.provider :linode do |provider|
-        provider.label = rbd_mirror.vm.hostname
-      end
-    end
-  end
-
-  (0..NISCSI_GWS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}iscsi_gw#{i}" do |iscsi_gw|
-      iscsi_gw.vm.hostname = "#{LABEL_PREFIX}ceph-iscsi-gw#{i}"
-      if ASSIGN_STATIC_IP
-        iscsi_gw.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.9#{i}"
-      end
-      # Virtualbox
-      iscsi_gw.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      iscsi_gw.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      iscsi_gw.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = true
-      end
-      # Parallels
-      iscsi_gw.vm.provider "parallels" do |prl|
-        prl.name = "ceph-iscsi-gw#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      iscsi_gw.vm.provider :linode do |provider|
-        provider.label = iscsi_gw.vm.hostname
-      end
-    end
-  end
-
-  (0..NMONS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}mon#{i}" do |mon|
-      mon.vm.hostname = "#{LABEL_PREFIX}ceph-mon#{i}"
-      if ASSIGN_STATIC_IP
-        mon.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.1#{i}"
-      end
-      # Virtualbox
-      mon.vm.provider :virtualbox do |vb|
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      mon.vm.provider :vmware_fusion do |v|
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      mon.vm.provider :libvirt do |lv|
-        lv.memory = MEMORY
-        lv.random_hostname = false
-      end
-
-      # Parallels
-      mon.vm.provider "parallels" do |prl|
-        prl.name = "ceph-mon#{i}"
-        prl.memory = "#{MEMORY}"
-      end
-
-      mon.vm.provider :linode do |provider|
-        provider.label = mon.vm.hostname
-      end
-    end
-  end
-
-  (0..NOSDS - 1).each do |i|
-    config.vm.define "#{LABEL_PREFIX}osd#{i}" do |osd|
-      osd.vm.hostname = "#{LABEL_PREFIX}ceph-osd#{i}"
-      if ASSIGN_STATIC_IP
-        osd.vm.network :private_network,
-          ip: "#{PUBLIC_SUBNET}.10#{i}"
-        osd.vm.network :private_network,
-          ip: "#{CLUSTER_SUBNET}.20#{i}"
-      end
-      # Virtualbox
-      osd.vm.provider :virtualbox do |vb|
-        # Create our own controller for consistency and to remove VM dependency
-        vb.customize ['storagectl', :id,
-                      '--name', 'OSD Controller',
-                      '--add', 'scsi']
-        (0..1).each do |d|
-          vb.customize ['createhd',
-                        '--filename', "disk-#{i}-#{d}",
-                        '--size', '11000'] unless File.exist?("disk-#{i}-#{d}.vdi")
-          vb.customize ['storageattach', :id,
-                        '--storagectl', 'OSD Controller',
-                        '--port', 3 + d,
-                        '--device', 0,
-                        '--type', 'hdd',
-                        '--medium', "disk-#{i}-#{d}.vdi"]
-        end
-        vb.customize ['modifyvm', :id, '--memory', "#{MEMORY}"]
-      end
-
-      # VMware
-      osd.vm.provider :vmware_fusion do |v|
-        (0..1).each do |d|
-          v.vmx["scsi0:#{d + 1}.present"] = 'TRUE'
-          v.vmx["scsi0:#{d + 1}.fileName"] =
-            create_vmdk("disk-#{i}-#{d}", '11000MB')
-        end
-        v.vmx['memsize'] = "#{MEMORY}"
-      end
-
-      # Libvirt
-      driverletters = ('a'..'z').to_a
-      osd.vm.provider :libvirt do |lv|
-        # always make /dev/sd{a/b/c} so that CI can ensure that
-        # virtualbox and libvirt will have the same devices to use for OSDs
-        (0..2).each do |d|
-          lv.storage :file, :device => "hd#{driverletters[d]}", :path => "disk-#{i}-#{d}-#{DISK_UUID}.disk", :size => '12G', :bus => "ide"
-        end
-        lv.memory = MEMORY
-        lv.random_hostname = false
-      end
-
-      # Parallels
-      osd.vm.provider "parallels" do |prl|
-        prl.name = "ceph-osd#{i}"
-        prl.memory = "#{MEMORY}"
-        (0..1).each do |d|
-          prl.customize ["set", :id,
-                         "--device-add",
-                         "hdd",
-                         "--iface",
-                         "sata"]
-        end
-      end
-
-      osd.vm.provider :linode do |provider|
-        provider.label = osd.vm.hostname
-      end
-
-      # Run the provisioner after the last machine comes up
-      osd.vm.provision 'ansible', &ansible_provision if i == (NOSDS - 1)
-    end
-  end
-end
new file mode 120000 (symlink)
index 0000000000000000000000000000000000000000..dfd7436c93f56089cc77635c8d49e0ccebcf0d3b
--- /dev/null
@@ -0,0 +1 @@
+../../../../../Vagrantfile
\ No newline at end of file
index 3a078510d695080a7261ce15b38d2c3ae9d654ea..663b7dc240942e0f60463961c584706eea21230a 100644 (file)
@@ -15,6 +15,8 @@ cluster_network: "192.168.16.0/24"
 journal_collocation: true
 ceph_rgw_civetweb_port: 8080
 ceph_osd_docker_devices: "{{ devices }}"
+ceph_osd_docker_prepare_env: -e CLUSTER={{ cluster }} -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_FORCE_ZAP=1
+ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE
 devices:
   - /dev/sda
   - /dev/sdb
index c8ddf66cbda474a39fa6afe1568b2d087348eaf8..ac0c1211607c37c2565594fe32eddb8279f8cbd7 100644 (file)
@@ -6,6 +6,7 @@ public_network: "192.168.3.0/24"
 cluster_network: "192.168.4.0/24"
 monitor_interface: eth1
 journal_size: 100
+osd_objectstore: "filestore"
 devices:
   - '/dev/sda'
   - '/dev/sdb'
diff --git a/tox.ini b/tox.ini
index ed6d87e7e1a8927064657a392923507580acfa55..db6a35c9a1c43f58c2f4ff2a16d6d8bafa9b8e24 100644 (file)
--- a/tox.ini
+++ b/tox.ini
@@ -1,5 +1,7 @@
 [tox]
-envlist = {jewel,kraken,luminous,rhcs}-{ansible2.2}-{xenial_cluster,journal_collocation,centos7_cluster,dmcrypt_journal,dmcrypt_journal_collocation,docker_cluster,purge_cluster,purge_dmcrypt,docker_dedicated_journal,docker_dmcrypt_journal_collocation,update_dmcrypt,update_cluster,cluster,purge_docker_cluster,update_docker_cluster,bluestore_journal_collocation,bluestore_cluster}
+envlist = {jewel,luminous,rhcs}-{ansible2.2}-{xenial_cluster,journal_collocation,centos7_cluster,dmcrypt_journal,dmcrypt_journal_collocation,docker_cluster,purge_cluster,purge_dmcrypt,docker_dedicated_journal,docker_dmcrypt_journal_collocation,update_dmcrypt,update_cluster,cluster,purge_docker_cluster,update_docker_cluster}
+  {luminous}-{ansible2.2}-{bluestore_journal_collocation,bluestore_cluster,bluestore_dmcrypt_journal,bluestore_dmcrypt_journal_collocation,bluestore_docker_cluster,bluestore_docker_dedicated_journal,bluestore_docker_dmcrypt_journal_collocation}
+
 skipsdist = True
 
 # extra commands for purging clusters
@@ -64,12 +66,16 @@ setenv=
   purge_docker_cluster: PURGE_PLAYBOOK = purge-docker-cluster.yml
   docker_dedicated_journal: PLAYBOOK = site-docker.yml.sample
   docker_dmcrypt_journal_collocation: PLAYBOOK = site-docker.yml.sample
+  bluestore_docker_cluster: PLAYBOOK = site-docker.yml.sample
+  bluestore_docker_dedicated_journal: PLAYBOOK = site-docker.yml.sample
+  bluestore_docker_dmcrypt_journal_collocation: PLAYBOOK = site-docker.yml.sample
+
   rhcs: CEPH_STABLE_RELEASE = jewel
   jewel: CEPH_STABLE_RELEASE = jewel
   jewel: CEPH_DOCKER_IMAGE_TAG = tag-build-master-jewel-ubuntu-16.04
-  kraken: CEPH_STABLE_RELEASE = kraken
-  kraken: UPDATE_CEPH_STABLE_RELEASE = luminous
+  jewel: UPDATE_CEPH_STABLE_RELEASE = kraken
   luminous: CEPH_STABLE_RELEASE = luminous
+  luminous: CEPH_DOCKER_IMAGE_TAG = tag-build-master-luminous-ubuntu-16.04
   luminous: UPDATE_CEPH_STABLE_RELEASE = luminous
   luminous: UPDATE_CEPH_DOCKER_IMAGE_TAG = tag-build-master-luminous-ubuntu-16.04
 deps=
@@ -102,6 +108,11 @@ changedir=
   update_cluster: {toxinidir}/tests/functional/centos/7/cluster
   bluestore_journal_collocation: {toxinidir}/tests/functional/centos/7/bs-journal-collocation
   bluestore_cluster: {toxinidir}/tests/functional/centos/7/bs-cluster
+  bluestore_dmcrypt_journal: {toxinidir}/tests/functional/centos/7/bs-dmcrypt-dedicated-journal
+  bluestore_dmcrypt_journal_collocation: {toxinidir}/tests/functional/centos/7/bs-dmcrypt-journal-colloc
+  bluestore_docker_cluster: {toxinidir}/tests/functional/centos/7/bs-docker-cluster
+  bluestore_docker_dedicated_journal: {toxinidir}/tests/functional/centos/7/bs-docker-cluster-dedicated-journal
+  bluestore_docker_dmcrypt_journal_collocation: {toxinidir}/tests/functional/centos/7/bs-docker-dmcrypt-journal-colloc
 
 commands=
   rhcs: ansible-playbook -vv -i "localhost," -c local {toxinidir}/tests/functional/rhcs_setup.yml --extra-vars "change_dir={changedir}" --tags "vagrant_setup"