crimson::osd::ShardServices& shard_services,
const ec_profile_t&,
uint64_t)
- : PGBackend{shard, coll, &shard_services.get_store()}
+ : PGBackend{shard, coll, shard_services}
{
// todo
}
PGBackend::PGBackend(shard_id_t shard,
CollectionRef coll,
- crimson::os::FuturizedStore* store)
+ crimson::osd::ShardServices &shard_services)
: shard{shard},
coll{coll},
- store{store}
+ shard_services{shard_services},
+ store{&shard_services.get_store()}
{}
PGBackend::load_metadata_iertr::future
using rep_op_fut_t =
std::tuple<interruptible_future<>,
interruptible_future<crimson::osd::acked_peers_t>>;
- PGBackend(shard_id_t shard, CollectionRef coll, crimson::os::FuturizedStore* store);
+ PGBackend(shard_id_t shard, CollectionRef coll,
+ crimson::osd::ShardServices &shard_services);
virtual ~PGBackend() = default;
static std::unique_ptr<PGBackend> create(pg_t pgid,
const pg_shard_t pg_shard,
protected:
const shard_id_t shard;
CollectionRef coll;
+ crimson::osd::ShardServices &shard_services;
crimson::os::FuturizedStore* store;
bool stopping = false;
std::optional<peering_info_t> peering;
pg_shard_t whoami,
ReplicatedBackend::CollectionRef coll,
crimson::osd::ShardServices& shard_services)
- : PGBackend{whoami.shard, coll, &shard_services.get_store()},
+ : PGBackend{whoami.shard, coll, shard_services},
pgid{pgid},
- whoami{whoami},
- shard_services{shard_services}
+ whoami{whoami}
{}
ReplicatedBackend::ll_read_ierrorator::future<ceph::bufferlist>
std::vector<pg_log_entry_t>&& log_entries) final;
const pg_t pgid;
const pg_shard_t whoami;
- crimson::osd::ShardServices& shard_services;
ceph_tid_t next_txn_id = 0;
class pending_on_t : public seastar::weakly_referencable<pending_on_t> {
public: