/*
* pg lock may or may not be held
*/
-void PG::_scan_list(ScrubMap &map, vector<hobject_t> &ls, bool deep)
+void PG::_scan_list(
+ ScrubMap &map, vector<hobject_t> &ls, bool deep,
+ ThreadPool::TPHandle &handle)
{
dout(10) << "_scan_list scanning " << ls.size() << " objects"
<< (deep ? " deeply" : "") << dendl;
* build a scrub map over a chunk without releasing the lock
* only used by chunky scrub
*/
-int PG::build_scrub_map_chunk(ScrubMap &map,
- hobject_t start, hobject_t end, bool deep)
+int PG::build_scrub_map_chunk(
+ ScrubMap &map,
+ hobject_t start, hobject_t end, bool deep,
+ ThreadPool::TPHandle &handle)
{
dout(10) << "build_scrub_map" << dendl;
dout(20) << "scrub_map_chunk [" << start << "," << end << ")" << dendl;
return ret;
}
- _scan_list(map, ls, deep);
+ _scan_list(map, ls, deep, handle);
// pg attrs
osd->store->collection_getattrs(coll, map.attrs);
* build a (sorted) summary of pg content for purposes of scrubbing
* called while holding pg lock
*/
-void PG::build_scrub_map(ScrubMap &map)
+void PG::build_scrub_map(ScrubMap &map, ThreadPool::TPHandle &handle)
{
dout(10) << "build_scrub_map" << dendl;
vector<hobject_t> ls;
osd->store->collection_list(coll, ls);
- _scan_list(map, ls, false);
+ _scan_list(map, ls, false, handle);
lock();
if (epoch != info.history.same_interval_since) {
* build a summary of pg content changed starting after v
* called while holding pg lock
*/
-void PG::build_inc_scrub_map(ScrubMap &map, eversion_t v)
+void PG::build_inc_scrub_map(
+ ScrubMap &map, eversion_t v,
+ ThreadPool::TPHandle &handle)
{
map.valid_through = last_update_applied;
map.incr_since = v;
}
}
- _scan_list(map, ls, false);
+ _scan_list(map, ls, false, handle);
// pg attrs
osd->store->collection_getattrs(coll, map.attrs);
* for pushes to complete in case of recent recovery. Build a single
* scrubmap of objects that are in the range [msg->start, msg->end).
*/
-void PG::replica_scrub(MOSDRepScrub *msg)
+void PG::replica_scrub(
+ MOSDRepScrub *msg,
+ ThreadPool::TPHandle &handle)
{
assert(!scrubber.active_rep_scrub);
dout(7) << "replica_scrub" << dendl;
return;
}
- build_scrub_map_chunk(map, msg->start, msg->end, msg->deep);
+ build_scrub_map_chunk(
+ map, msg->start, msg->end, msg->deep,
+ handle);
} else {
if (msg->scrub_from > eversion_t()) {
return;
}
}
- build_inc_scrub_map(map, msg->scrub_from);
+ build_inc_scrub_map(map, msg->scrub_from, handle);
scrubber.finalizing = 0;
} else {
- build_scrub_map(map);
+ build_scrub_map(map, handle);
}
if (msg->map_epoch < info.history.same_interval_since) {
* scrub will be chunky if all OSDs in PG support chunky scrub
* scrub will fall back to classic in any other case
*/
-void PG::scrub()
+void PG::scrub(ThreadPool::TPHandle &handle)
{
lock();
if (deleting) {
}
if (scrubber.is_chunky) {
- chunky_scrub();
+ chunky_scrub(handle);
} else {
- classic_scrub();
+ classic_scrub(handle);
}
unlock();
* Flag set when we're in the finalize stage.
*
*/
-void PG::classic_scrub()
+void PG::classic_scrub(ThreadPool::TPHandle &handle)
{
if (!scrubber.active) {
dout(10) << "scrub start" << dendl;
// Unlocks and relocks...
scrubber.primary_scrubmap = ScrubMap();
- build_scrub_map(scrubber.primary_scrubmap);
+ build_scrub_map(scrubber.primary_scrubmap, handle);
if (scrubber.epoch_start != info.history.same_interval_since) {
dout(10) << "scrub pg changed, aborting" << dendl;
if (scrubber.primary_scrubmap.valid_through != log.head) {
ScrubMap incr;
- build_inc_scrub_map(incr, scrubber.primary_scrubmap.valid_through);
+ build_inc_scrub_map(incr, scrubber.primary_scrubmap.valid_through, handle);
scrubber.primary_scrubmap.merge_incr(incr);
}
* scrubber.state encodes the current state of the scrub (refer to state diagram
* for details).
*/
-void PG::chunky_scrub() {
+void PG::chunky_scrub(ThreadPool::TPHandle &handle) {
// check for map changes
if (scrubber.is_chunky_scrub_active()) {
if (scrubber.epoch_start != info.history.same_interval_since) {
// build my own scrub map
ret = build_scrub_map_chunk(scrubber.primary_scrubmap,
scrubber.start, scrubber.end,
- scrubber.deep);
+ scrubber.deep,
+ handle);
if (ret < 0) {
dout(5) << "error building scrub map: " << ret << ", aborting" << dendl;
scrub_clear_state();
#include "msg/Messenger.h"
#include "messages/MOSDRepScrub.h"
#include "messages/MOSDPGLog.h"
+#include "common/WorkQueue.h"
#include "common/DecayCounter.h"
map<hobject_t, int> &authoritative,
map<hobject_t, set<int> > &inconsistent_snapcolls,
ostream &errorstream);
- void scrub();
- void classic_scrub();
- void chunky_scrub();
+ void scrub(ThreadPool::TPHandle &handle);
+ void classic_scrub(ThreadPool::TPHandle &handle);
+ void chunky_scrub(ThreadPool::TPHandle &handle);
void scrub_compare_maps();
void scrub_process_inconsistent();
void scrub_finalize();
void scrub_finish();
void scrub_clear_state();
bool scrub_gather_replica_maps();
- void _scan_list(ScrubMap &map, vector<hobject_t> &ls, bool deep);
+ void _scan_list(
+ ScrubMap &map, vector<hobject_t> &ls, bool deep,
+ ThreadPool::TPHandle &handle);
void _request_scrub_map_classic(int replica, eversion_t version);
void _request_scrub_map(int replica, eversion_t version,
hobject_t start, hobject_t end, bool deep);
- int build_scrub_map_chunk(ScrubMap &map,
- hobject_t start, hobject_t end, bool deep);
- void build_scrub_map(ScrubMap &map);
- void build_inc_scrub_map(ScrubMap &map, eversion_t v);
+ int build_scrub_map_chunk(
+ ScrubMap &map,
+ hobject_t start, hobject_t end, bool deep,
+ ThreadPool::TPHandle &handle);
+ void build_scrub_map(ScrubMap &map, ThreadPool::TPHandle &handle);
+ void build_inc_scrub_map(
+ ScrubMap &map, eversion_t v, ThreadPool::TPHandle &handle);
virtual void _scrub(ScrubMap &map) { }
virtual void _scrub_clear_state() { }
virtual void _scrub_finish() { }
void reg_next_scrub();
void unreg_next_scrub();
- void replica_scrub(class MOSDRepScrub *op);
+ void replica_scrub(
+ class MOSDRepScrub *op,
+ ThreadPool::TPHandle &handle);
void sub_op_scrub_map(OpRequestRef op);
void sub_op_scrub_reserve(OpRequestRef op);
void sub_op_scrub_reserve_reply(OpRequestRef op);