When a new osd joins a cluster, all pgs containing it must eventually backfill
to it. If all of these backfills happen simultaneously, it would put excessive
-load on the osd. osd_num_concurrent_backfills limits the number of outgoing or
-incoming backfills on a single node.
+load on the osd. osd_max_backfills limits the number of outgoing or
+incoming backfills on a single node. The maximum number of outgoing backfills is
+osd_max_backfills. The maximum number of incoming backfills is
+osd_max_backfills. Therefore there can be a maximum of osd_max_backfills * 2
+simultaneous backfills on one osd.
Each OSDService now has two AsyncReserver instances: one for backfills going
from the osd (local_reserver) and one for backfills going to the osd
It's important that we always grab the local reservation before the remote
reservation in order to prevent a circular dependency.
-We want to minimize the risk of data loss by prioritizing the order in which
-PGs are recovered. We use 3 AsyncReserver priorities to hand out reservations.
-The highest priority is log based recovery (RECOVERY) since this must always
-complete before backfill can start. The next priority is backfill of degraded
-PGs (BACKFILL_HIGH). The lowest priority is backfill of non-degraded PGs
-(BACKFILL_LOW).
+We want to minimize the risk of data loss by prioritizing the order in
+which PGs are recovered. The highest priority is log based recovery
+(OSD_RECOVERY_PRIORITY_MAX) since this must always complete before
+backfill can start. The next priority is backfill of degraded PGs and
+is a function of the degradation. A backfill for a PG missing two
+replicas will have a priority higher than a backfill for a PG missing
+one replica. The lowest priority is backfill of non-degraded PGs.