From d7beff636c4a9f306c542729f728bbf962e284a9 Mon Sep 17 00:00:00 2001 From: Patrick Donnelly Date: Fri, 26 Feb 2021 09:13:21 -0800 Subject: [PATCH] qa: delete all fs during tearDown During the ceph task Unwind, the MDS are stopped. If any file system still exists, we will see failover messages in the cluster log. Fixes: https://tracker.ceph.com/issues/49510 Signed-off-by: Patrick Donnelly (cherry picked from commit 4e6d9e6a0484357d483fb1a183d1bfbcc33b021a) Conflicts: qa/tasks/cephfs/cephfs_test_case.py - trivial resolution --- qa/tasks/cephfs/cephfs_test_case.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/qa/tasks/cephfs/cephfs_test_case.py b/qa/tasks/cephfs/cephfs_test_case.py index f901f44ba494..8136d4d5a65b 100644 --- a/qa/tasks/cephfs/cephfs_test_case.py +++ b/qa/tasks/cephfs/cephfs_test_case.py @@ -176,6 +176,9 @@ class CephFSTestCase(CephTestCase): for m in self.mounts: m.teardown() + # To prevent failover messages during Unwind of ceph task + self.mds_cluster.delete_all_filesystems() + for i, m in enumerate(self.mounts): m.client_id = self._original_client_ids[i] -- 2.47.3