]> git-server-git.apps.pok.os.sepia.ceph.com Git - ceph.git/commitdiff
osd/PG: kill PG_STATE_REPLAY 13942/head
authorxie xingguo <xie.xingguo@zte.com.cn>
Mon, 20 Mar 2017 02:17:37 +0000 (10:17 +0800)
committerxie xingguo <xie.xingguo@zte.com.cn>
Mon, 20 Mar 2017 02:24:25 +0000 (10:24 +0800)
This replay state of PG is not functionable and
hence should not be made visible to user.

Signed-off-by: xie xingguo <xie.xingguo@zte.com.cn>
doc/rados/operations/pg-states.rst
src/mon/MonCommands.h
src/osd/PG.cc
src/osd/osd_types.cc
src/osd/osd_types.h

index 940885469433bb63f1b6741bb1314bc6d21df993..03a92426de5f923f4105d1064fe6436829d8532e 100644 (file)
@@ -19,9 +19,6 @@ map is ``active + clean``.
 *Down*
   A replica with necessary data is down, so the placement group is offline.
 
-*Replay*
-  The placement group is waiting for clients to replay operations after an OSD crashed.
-
 *Scrubbing*
   Ceph is checking the placement group for inconsistencies.
 
index cefb2102cfef90d1474dda20c9f556436aeb1f5e..7589054bfa996224aea7d9ef04fe27e9967d3c6a 100644 (file)
@@ -139,21 +139,21 @@ COMMAND("pg dump_stuck " \
        "pg", "r", "cli,rest")
 COMMAND("pg ls-by-pool " \
         "name=poolstr,type=CephString " \
-       "name=states,type=CephChoices,strings=active|clean|down|replay|scrubbing|degraded|inconsistent|peering|repair|recovering|backfill_wait|incomplete|stale|remapped|deep_scrub|backfill|backfill_toofull|recovery_wait|undersized|activating|peered,n=N,req=false ", \
+       "name=states,type=CephChoices,strings=active|clean|down|scrubbing|degraded|inconsistent|peering|repair|recovering|backfill_wait|incomplete|stale|remapped|deep_scrub|backfill|backfill_toofull|recovery_wait|undersized|activating|peered,n=N,req=false ", \
        "list pg with pool = [poolname]", "pg", "r", "cli,rest")
 COMMAND("pg ls-by-primary " \
         "name=osd,type=CephOsdName " \
         "name=pool,type=CephInt,req=false " \
-       "name=states,type=CephChoices,strings=active|clean|down|replay|scrubbing|degraded|inconsistent|peering|repair|recovering|backfill_wait|incomplete|stale|remapped|deep_scrub|backfill|backfill_toofull|recovery_wait|undersized|activating|peered,n=N,req=false ", \
+       "name=states,type=CephChoices,strings=active|clean|down|scrubbing|degraded|inconsistent|peering|repair|recovering|backfill_wait|incomplete|stale|remapped|deep_scrub|backfill|backfill_toofull|recovery_wait|undersized|activating|peered,n=N,req=false ", \
        "list pg with primary = [osd]", "pg", "r", "cli,rest")
 COMMAND("pg ls-by-osd " \
         "name=osd,type=CephOsdName " \
         "name=pool,type=CephInt,req=false " \
-       "name=states,type=CephChoices,strings=active|clean|down|replay|scrubbing|degraded|inconsistent|peering|repair|recovering|backfill_wait|incomplete|stale|remapped|deep_scrub|backfill|backfill_toofull|recovery_wait|undersized|activating|peered,n=N,req=false ", \
+       "name=states,type=CephChoices,strings=active|clean|down|scrubbing|degraded|inconsistent|peering|repair|recovering|backfill_wait|incomplete|stale|remapped|deep_scrub|backfill|backfill_toofull|recovery_wait|undersized|activating|peered,n=N,req=false ", \
        "list pg on osd [osd]", "pg", "r", "cli,rest")
 COMMAND("pg ls " \
         "name=pool,type=CephInt,req=false " \
-       "name=states,type=CephChoices,strings=active|clean|down|replay|scrubbing|degraded|inconsistent|peering|repair|recovering|backfill_wait|incomplete|stale|remapped|deep_scrub|backfill|backfill_toofull|recovery_wait|undersized|activating|peered,n=N,req=false ", \
+       "name=states,type=CephChoices,strings=active|clean|down|scrubbing|degraded|inconsistent|peering|repair|recovering|backfill_wait|incomplete|stale|remapped|deep_scrub|backfill|backfill_toofull|recovery_wait|undersized|activating|peered,n=N,req=false ", \
        "list pg with specific pool, osd, state", "pg", "r", "cli,rest")
 COMMAND("pg map name=pgid,type=CephPgid", "show mapping of pg to osds", \
        "pg", "r", "cli,rest")
index 44ef72f02109122620bed951f452c0bf37f3c02c..bc03bb5c1339f2e7d08d4f137ee3f2d64b88fbc1 100644 (file)
@@ -7071,7 +7071,6 @@ void PG::RecoveryState::Active::exit()
   pg->state_clear(PG_STATE_BACKFILL_TOOFULL);
   pg->state_clear(PG_STATE_BACKFILL_WAIT);
   pg->state_clear(PG_STATE_RECOVERY_WAIT);
-  pg->state_clear(PG_STATE_REPLAY);
   utime_t dur = ceph_clock_now() - enter_time;
   pg->osd->recoverystate_perf->tinc(rs_active_latency, dur);
   pg->agent_stop();
index c7671356a02d7474f4cb74c92461dd57194282c7..7eedcd3a2e456d1283c95f7e7790ed3ff8320f33 100644 (file)
@@ -793,8 +793,6 @@ std::string pg_state_string(int state)
     oss << "recovering+";
   if (state & PG_STATE_DOWN)
     oss << "down+";
-  if (state & PG_STATE_REPLAY)
-    oss << "replay+";
   if (state & PG_STATE_UNDERSIZED)
     oss << "undersized+";
   if (state & PG_STATE_DEGRADED)
@@ -843,8 +841,6 @@ int pg_string_state(const std::string& state)
     type = PG_STATE_CLEAN;
   else if (state == "down")
     type = PG_STATE_DOWN;
-  else if (state == "replay")
-    type = PG_STATE_REPLAY;
   else if (state == "scrubbing")
     type = PG_STATE_SCRUBBING;
   else if (state == "degraded")
index 7fe56fe707f8378fafb09d0235ef10aae74773c3..7e759884d1850dd254741874e41974d4ff92c55b 100644 (file)
@@ -954,7 +954,7 @@ inline ostream& operator<<(ostream& out, const osd_stat_t& s) {
 #define PG_STATE_ACTIVE       (1<<1)  // i am active.  (primary: replicas too)
 #define PG_STATE_CLEAN        (1<<2)  // peers are complete, clean of stray replicas.
 #define PG_STATE_DOWN         (1<<4)  // a needed replica is down, PG offline
-#define PG_STATE_REPLAY       (1<<5)  // crashed, waiting for replay
+//#define PG_STATE_REPLAY       (1<<5)  // crashed, waiting for replay
 //#define PG_STATE_STRAY      (1<<6)  // i must notify the primary i exist.
 //#define PG_STATE_SPLITTING    (1<<7)  // i am splitting
 #define PG_STATE_SCRUBBING    (1<<8)  // scrubbing