From 92b49094e78bdd52f9e474e362c8710c4fc8c0e6 Mon Sep 17 00:00:00 2001 From: Sage Weil Date: Wed, 12 Feb 2020 10:56:22 -0600 Subject: [PATCH] cephadm: avoid trigger old podman bug This ticket seems to suggest that (1) the root cause is related to an exec that is orphaned and screws up the container state (due to, e.g., ssh dropping, or a timeout), (2) -f may be needed, sometimes, to recover, and (3) newer versions fix it. https://github.com/containers/libpod/issues/3226 Way back in 26f9fe54cb635cbcd8f74849d6fa3528cdf5d755 we found that using -f the first time around was a Bad Idea, so we'd rather avoid this. Instead, just avoid triggering the bug. Signed-off-by: Sage Weil --- qa/workunits/cephadm/test_cephadm.sh | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/qa/workunits/cephadm/test_cephadm.sh b/qa/workunits/cephadm/test_cephadm.sh index 8676ddced52..c47e694328b 100755 --- a/qa/workunits/cephadm/test_cephadm.sh +++ b/qa/workunits/cephadm/test_cephadm.sh @@ -313,7 +313,8 @@ $CEPHADM enter --fsid $FSID --name mgr.x -- test -d /var/lib/ceph/mgr/ceph-x $CEPHADM enter --fsid $FSID --name mon.a -- pidof ceph-mon expect_false $CEPHADM enter --fsid $FSID --name mgr.x -- pidof ceph-mon $CEPHADM enter --fsid $FSID --name mgr.x -- pidof ceph-mgr -expect_false $CEPHADM --timeout 1 enter --fsid $FSID --name mon.a -- sleep 10 +# this triggers a bug in older versions of podman, including 18.04's 1.6.2 +#expect_false $CEPHADM --timeout 1 enter --fsid $FSID --name mon.a -- sleep 10 $CEPHADM --timeout 10 enter --fsid $FSID --name mon.a -- sleep 1 ## ceph-volume -- 2.39.5