]> git.apps.os.sepia.ceph.com Git - ceph.git/commitdiff
osd/PeeringState: do not exclude up from acting_recovery_backfill 32064/head
authorxie xingguo <xie.xingguo@zte.com.cn>
Wed, 20 Nov 2019 00:31:54 +0000 (08:31 +0800)
committerNathan Cutler <ncutler@suse.com>
Fri, 6 Dec 2019 12:55:13 +0000 (13:55 +0100)
If we choose a primary that does not belong to the current up set,
and all up peers are still recoverable, then we might end up excluding
some up peer from the acting_recovery_backfill set too due to the
"want size <= pool size" constraint (since https://github.com/ceph/ceph/pull/24035),
as a result of which all up peers might not get recovered in one go.

Fix by falling through any oversized want set to async recovery, which
should be able to handle it nicely.

Fixes: https://tracker.ceph.com/issues/42577
Signed-off-by: xie xingguo <xie.xingguo@zte.com.cn>
(cherry picked from commit 22c8cdad8ee1d7376c7d200bdb6ec94ed6d3b5e6)

Conflicts:
    src/osd/PeeringState.cc
- file does not exist in nautilus; made changes manually in src/osd/PG.cc
- used dout(10) instead of psdout(10)

src/osd/PG.cc

index a103c01c0eacdd4f9be50860ecb8897a44535f42..25fd932bd7cbde7c3ae9927dfed7cf3d367f4b0d 100644 (file)
@@ -1788,6 +1788,14 @@ bool PG::choose_acting(pg_shard_t &auth_log_shard_id,
       choose_async_recovery_replicated(all_info, auth_log_shard->second, &want, &want_async_recovery, get_osdmap());
     }
   }
+  while (want.size() > pool.info.size) {
+    // async recovery should have taken out as many osds as it can.
+    // if not, then always evict the last peer
+    // (will get synchronously recovered later)
+    dout(10) << __func__ << " evicting osd." << want.back()
+               << " from oversized want " << want << dendl;
+    want.pop_back();
+  }
   if (want != acting) {
     dout(10) << __func__ << " want " << want << " != acting " << acting
             << ", requesting pg_temp change" << dendl;