From: Patrick Donnelly Date: Mon, 11 Mar 2019 16:10:57 +0000 (-0700) Subject: qa: unmount clients prior to marking fs down X-Git-Tag: v14.2.0~48^2 X-Git-Url: http://git-server-git.apps.pok.os.sepia.ceph.com/?a=commitdiff_plain;h=refs%2Fpull%2F26893%2Fhead;p=ceph.git qa: unmount clients prior to marking fs down Evicted RHEL7.5 clients may hang. Fixes: http://tracker.ceph.com/issues/38677 Signed-off-by: Patrick Donnelly --- diff --git a/qa/tasks/cephfs/test_failover.py b/qa/tasks/cephfs/test_failover.py index 63bfce6e8c2b..f1bb0f22a5b3 100644 --- a/qa/tasks/cephfs/test_failover.py +++ b/qa/tasks/cephfs/test_failover.py @@ -96,6 +96,8 @@ class TestClusterResize(CephFSTestCase): That marking a FS down does not generate a health warning """ + self.mount_a.umount_wait() + self.fs.set_down() try: self.wait_for_health("", 30) @@ -111,6 +113,8 @@ class TestClusterResize(CephFSTestCase): That marking a FS down twice does not wipe old_max_mds. """ + self.mount_a.umount_wait() + self.grow(2) self.fs.set_down() self.fs.wait_for_daemons() @@ -123,6 +127,8 @@ class TestClusterResize(CephFSTestCase): That setting max_mds undoes down. """ + self.mount_a.umount_wait() + self.fs.set_down() self.fs.wait_for_daemons() self.grow(2) @@ -133,6 +139,8 @@ class TestClusterResize(CephFSTestCase): That down setting toggles and sets max_mds appropriately. """ + self.mount_a.umount_wait() + self.fs.set_down() self.fs.wait_for_daemons() self.assertEqual(self.fs.get_var("max_mds"), 0)