From: Kefu Chai Date: Wed, 10 May 2017 11:42:08 +0000 (+0800) Subject: doc: add descriptions for mon/mgr options X-Git-Tag: v12.1.0~10^2~98^2 X-Git-Url: http://git-server-git.apps.pok.os.sepia.ceph.com/?a=commitdiff_plain;h=4b52a03859e102c7948729fea0c474bbd45740ab;p=ceph.git doc: add descriptions for mon/mgr options Signed-off-by: Kefu Chai --- diff --git a/doc/install/manual-deployment.rst b/doc/install/manual-deployment.rst index cf14d4b838a6..5e58f57c0e12 100644 --- a/doc/install/manual-deployment.rst +++ b/doc/install/manual-deployment.rst @@ -58,7 +58,7 @@ a number of things: For example, when you run multiple clusters in a `federated architecture`_, the cluster name (e.g., ``us-west``, ``us-east``) identifies the cluster for the current CLI session. **Note:** To identify the cluster name on the - command line interface, specify the a Ceph configuration file with the + command line interface, specify the Ceph configuration file with the cluster name (e.g., ``ceph.conf``, ``us-west.conf``, ``us-east.conf``, etc.). Also see CLI usage (``ceph --cluster {cluster-name}``). diff --git a/doc/rados/configuration/mon-config-ref.rst b/doc/rados/configuration/mon-config-ref.rst index b19461f7a621..b27a8c8619c5 100644 --- a/doc/rados/configuration/mon-config-ref.rst +++ b/doc/rados/configuration/mon-config-ref.rst @@ -106,6 +106,11 @@ A consensus requires a majority of monitors running to establish a quorum for consensus about the cluster map (e.g., 1; 2 out of 3; 3 out of 5; 4 out of 6; etc.). +``mon force quorum join`` + +:Description: Force monitor to join quorum even if it has been previously removed from the map +:Type: Boolean +:Default: ``False`` .. index:: Ceph Monitor; consistency @@ -314,6 +319,126 @@ by setting it in the ``[mon]`` section of the configuration file. :Default: ``/var/lib/ceph/mon/$cluster-$id`` +``mon data size warn`` + +:Description: Issue a ``HEALTH_WARN`` in cluster log when the monitor's data + store goes over 15GB. +:Type: Integer +:Default: 15*1024*1024*1024* + + +``mon data avail warn`` + +:Description: Issue a ``HEALTH_WARN`` in cluster log when the available disk + space of monitor's data store is lower or equal to this + percentage. +:Type: Integer +:Default: 30 + + +``mon data avail crit`` + +:Description: Issue a ``HEALTH_ERR`` in cluster log when the available disk + space of monitor's data store is lower or equal to this + percentage. +:Type: Integer +:Default: 5 + + +``mon warn on cache pools without hit sets`` + +:Description: Issue a ``HEALTH_WARN`` in cluster log if a cache pool does not + have the hitset type set set. + See `hit set type <../operations/pools#hit-set-type>`_ for more + details. +:Type: Boolean +:Default: True + + +``mon warn on crush straw calc version zero`` + +:Description: Issue a ``HEALTH_WARN`` in cluster log if the CRUSH's + ``straw_calc_version`` is zero. See + `CRUSH map tunables <../operations/crush-map#tunables>`_ for + details. +:Type: Boolean +:Default: True + + +``mon warn on legacy crush tunables`` + +:Description: Issue a ``HEALTH_WARN`` in cluster log if + CRUSH tunables are too old (older than ``mon_min_crush_required_version``) +:Type: Boolean +:Default: True + + +``mon crush min required version`` + +:Description: The minimum tunable profile version required by the cluster. + See + `CRUSH map tunables <../operations/crush-map#tunables>`_ for + details. +:Type: String +:Default: ``firefly`` + + +``mon warn on osd down out interval zero`` + +:Description: Issue a ``HEALTH_WARN`` in cluster log if + ``mon osd down out interval`` is zero. Having this option set to + zero on the leader acts much like the ``noout`` flag. It's hard + to figure out what's going wrong with clusters witout the + ``noout`` flag set but acting like that just the same, so we + report a warning in this case. +:Type: Boolean +:Default: True + + +``mon cache target full warn ratio`` + +:Description: Position between pool's ``cache_target_full`` and + ``target_max_object`` where we start warning +:Type: Float +:Default: ``0.66`` + + +``mon health data update interval`` + +:Description: How often (in seconds) the monitor in quorum shares its health + status with its peers. (negative number disables it) +:Type: Float +:Default: ``60`` + + +``mon health to clog`` + +:Description: Enable sending health summary to cluster log periodically. +:Type: Boolean +:Default: True + + +``mon health to clog tick interval`` + +:Description: How often (in seconds) the monitor send health summary to cluster + log (a non-positive number disables it). If current health summary + is empty or identical to the last time, monitor will not send it + to cluster log. +:Type: Integer +:Default: 3600 + + +``mon health to clog interval`` + +:Description: How often (in seconds) the monitor send health summary to cluster + log (a non-positive number disables it). Monitor will always + send the summary to cluster log no matter if the summary changes + or not. +:Type: Integer +:Default: 60 + + + .. index:: Ceph Storage Cluster; capacity planning, Ceph Monitor; capacity planning Storage Capacity @@ -546,7 +671,9 @@ Trimming requires that the placement groups are ``active + clean``. ``mon sync timeout`` -:Description: +:Description: Number of seconds the monitor will wait for the next update + message from its sync provider before it gives up and bootstrap + again. :Type: Double :Default: ``30.0`` @@ -560,39 +687,123 @@ Trimming requires that the placement groups are ``active + clean``. ``mon sync max payload size`` -:Description: The maximum size for a sync payload. +:Description: The maximum size for a sync payload (in bytes). :Type: 32-bit Integer :Default: ``1045676`` -``mon accept timeout`` +``paxos max join drift`` -:Description: Number of seconds the Leader will wait for the Requester(s) to - accept a Paxos update. It is also used during the Paxos recovery - phase for similar purposes. +:Description: The maximum Paxos iterations before we must first sync the + monitor data stores. When a monitor finds that its peer is too + far ahead of it, it will first sync with data stores before moving + on. +:Type: Integer +:Default: ``10`` -:Type: Float -:Default: ``10.0`` +``paxos stash full interval`` +:Description: How often (in commits) to stash a full copy of the PaxosService state. + Current this setting only affects ``mds``, ``mon``, ``auth`` and ``mgr`` + PaxosServices. +:Type: Integer +:Default: 25 ``paxos propose interval`` :Description: Gather updates for this time interval before proposing - a map update. - + a map update. :Type: Double :Default: ``1.0`` +``paxos min`` + +:Description: The minimum number of paxos states to keep around +:Type: Integer +:Default: 500 + + ``paxos min wait`` :Description: The minimum amount of time to gather updates after a period of inactivity. - :Type: Double :Default: ``0.05`` +``paxos trim min`` + +:Description: Number of extra proposals tolerated before trimming +:Type: Integer +:Default: 250 + + +``paxos trim max`` + +:Description: The maximum number of extra proposals to trim at a time +:Type: Integer +:Default: 500 + + +``paxos service trim min`` + +:Description: The minimum amount of versions to trigger a trim (0 disables it) +:Type: Integer +:Default: 250 + + +``paxos service trim max`` + +:Description: The maximum amount of versions to trim during a single proposal (0 disables it) +:Type: Integer +:Default: 500 + + +``mon max log epochs`` + +:Description: The maximum amount of log epochs to trim during a single proposal +:Type: Integer +:Default: 500 + + +``mon max pgmap epochs`` + +:Description: The maximum amount of pgmap epochs to trim during a single proposal +:Type: Integer +:Default: 500 + + +``mon mds force trim to`` + +:Description: Force monitor to trim mdsmaps to this point (0 disables it. + dangerous, use with care) +:Type: Integer +:Default: 0 + + +``mon osd force trim to`` + +:Description: Force monitor to trim osdmaps to this point, even if there is + PGs not clean at the specified epoch (0 disables it. dangerous, + use with care) +:Type: Integer +:Default: 0 + +``mon osd cache size`` + +:Description: The size of osdmaps cache, not to rely on underlying store's cache +:Type: Integer +:Default: 10 + + +``mon election timeout`` + +:Description: On election proposer, maximum waiting time for all ACKs in seconds. +:Type: Float +:Default: ``5`` + + ``mon lease`` :Description: The length (in seconds) of the lease on the monitor's versions. @@ -600,22 +811,30 @@ Trimming requires that the placement groups are ``active + clean``. :Default: ``5`` -``mon lease renew interval`` +``mon lease renew interval factor`` -:Description: The interval (in seconds) for the Leader to renew the other - monitor's leases. - +:Description: ``mon lease`` \* ``mon lease renew interval factor`` will be the + interval for the Leader to renew the other monitor's leases. The + factor should be less than ``1.0``. :Type: Float -:Default: ``3`` +:Default: ``0.6`` -``mon lease ack timeout`` +``mon lease ack timeout factor`` -:Description: The number of seconds the Leader will wait for the Providers to - acknowledge the lease extension. - +:Description: The Leader will wait ``mon lease`` \* ``mon lease ack timeout factor`` + for the Providers to acknowledge the lease extension. :Type: Float -:Default: ``10.0`` +:Default: ``2.0`` + + +``mon accept timeout factor`` + +:Description: The Leader will wait ``mon lease`` \* ``mon accept timeout factor`` + for the Requester(s) to accept a Paxos update. It is also used + during the Paxos recovery phase for similar purposes. +:Type: Float +:Default: ``2.0`` ``mon min osdmap epochs`` @@ -640,42 +859,6 @@ Trimming requires that the placement groups are ``active + clean``. - -Slurp ------ - -In Ceph version 0.58 and earlier, when a Paxos service drifts beyond a given -number of versions, Ceph triggers the `slurp` mechanism, which establishes a -connection with the quorum Leader and obtains every single version the Leader -has for every service that has drifted. In Ceph versions 0.59 and later, slurp -will not work, because there is a single Paxos instance for all services. - -.. deprecated:: 0.58 - -``paxos max join drift`` - -:Description: The maximum Paxos iterations before we must first sync the - monitor data stores. -:Type: Integer -:Default: ``10`` - - -``mon slurp timeout`` - -:Description: The number of seconds the monitor has to recover using slurp - before the process is aborted and the monitor bootstraps. - -:Type: Double -:Default: ``10.0`` - - -``mon slurp bytes`` - -:Description: Limits the slurp messages to the specified number of bytes. -:Type: 32-bit Integer -:Default: ``256 * 1024`` - - .. index:: Ceph Monitor; clock Clock @@ -739,12 +922,19 @@ acceptable values. ``mon timecheck interval`` :Description: The time check interval (clock drift check) in seconds - for the leader. + for the Leader. :Type: Float :Default: ``300.0`` +``mon timecheck skew interval`` + +:Description: The time check interval (clock drift check) in seconds when in + presence of a skew in seconds for the Leader. +:Type: Float +:Default: ``30.0`` + Client ------ @@ -897,6 +1087,129 @@ Miscellaneous :Default: ``0.5`` +``mon osd prime pg temp max time estimate`` + +:Description: Maximum estimate of time spent on each PG before we prime all PGs + in parallel. +:Type: Float +:Default: ``0.25`` + + +``mon osd allow primary affinity`` + +:Description: allow ``primary_affinity`` to be set in the osdmap. +:Type: Boolean +:Default: False + + +``mon osd pool ec fast read`` + +:Description: Whether turn on fast read on the pool or not. It will be used as + the default setting of newly created erasure pools if ``fast_read`` + is not specified at create time. +:Type: Boolean +:Default: False + + +``mon mds skip sanity`` + +:Description: Skip safety assertions on FSMap (in case of bugs where we want to + continue anyway). Monitor terminates if the FSMap sanity check + fails, but we can disable it by enabling this option. +:Type: Boolean +:Default: False + + +``mon max mdsmap epochs`` + +:Description: The maximum amount of mdsmap epochs to trim during a single proposal. +:Type: Integer +:Default: 500 + + +``mon config key max entry size`` + +:Description: The maximum size of config-key entry (in bytes) +:Type: Integer +:Default: 4096 + + +``mon scrub interval`` + +:Description: How often (in seconds) the monitor scrub its store by comparing + the stored checksums with the computed ones of all the stored + keys. +:Type: Integer +:Default: 3600*24 + + +``mon scrub max keys`` + +:Description: The maximum number of keys to scrub each time. +:Type: Integer +:Default: 100 + + +``mon compact on start`` + +:Description: Compact the database used as Ceph Monitor store on + ``ceph-mon`` start. A manual compaction helps to shrink the + monitor database and improve the performance of it if the regular + compaction fails to work. +:Type: Boolean +:Default: False + + +``mon compact on bootstrap`` + +:Description: Compact the database used as Ceph Monitor store on + on bootstrap. Monitor starts probing each other for creating + a quorum after bootstrap. If it times out before joining the + quorum, it will start over and bootstrap itself again. +:Type: Boolean +:Default: False + + +``mon compact on trim`` + +:Description: Compact a certain prefix (including paxos) when we trim its old states. +:Type: Boolean +:Default: True + + +``mon cpu threads`` + +:Description: Number of threads for performing CPU intensive work on monitor. +:Type: Boolean +:Default: True + + +``mon osd mapping pgs per chunk`` + +:Description: We calculate the mapping from placement group to OSDs in chunks. + This option specifies the number of placement groups per chunk. +:Type: Integer +:Default: 4096 + + +``mon osd max split count`` + +:Description: Largest number of PGs per "involved" OSD to let split create. + When we increase the ``pg_num`` of a pool, the placement groups + will be splitted on all OSDs serving that pool. We want to avoid + extreme multipliers on PG splits. +:Type: Integer +:Default: 300 + + +``mon session timeout`` + +:Description: Monitor will terminate inactive sessions stay idle over this + time limit. +:Type: Integer +:Default: 300 + + .. _Paxos: http://en.wikipedia.org/wiki/Paxos_(computer_science) .. _Monitor Keyrings: ../../../dev/mon-bootstrap#secret-keys diff --git a/doc/rados/configuration/mon-lookup-dns.rst b/doc/rados/configuration/mon-lookup-dns.rst index 0b0cb5699a9d..9aa1d37396e9 100644 --- a/doc/rados/configuration/mon-lookup-dns.rst +++ b/doc/rados/configuration/mon-lookup-dns.rst @@ -12,6 +12,13 @@ This allows for less configuration on clients and monitors. Using a DNS update c By default clients and daemons will look for the TCP service called *ceph-mon* which is configured by the *mon_dns_srv_name* configuration directive. + +``mon dns srv name`` + +:Description: the service name used querying the DNS for the monitor hosts/addresses +:Type: String +:Default: ``ceph-mon`` + Example ------- When the DNS search domain is set to *example.com* a DNS zone file might contain the following elements. diff --git a/doc/rados/configuration/mon-osd-interaction.rst b/doc/rados/configuration/mon-osd-interaction.rst index c71b400ed0d0..ab57cb069a88 100644 --- a/doc/rados/configuration/mon-osd-interaction.rst +++ b/doc/rados/configuration/mon-osd-interaction.rst @@ -66,39 +66,46 @@ or by setting the value at runtime. |----+ Mark | | | OSD 2 | |<---+ Down | - + .. index:: OSD down report OSDs Report Down OSDs ===================== -By default, a Ceph OSD Daemon must report to the Ceph Monitors that another Ceph -OSD Daemon is ``down`` three times before the Ceph Monitors acknowledge that the -reported Ceph OSD Daemon is ``down``. By default, only one -Ceph OSD Daemon is required to report another Ceph OSD Daemon ``down``. You can -change the number of Ceph OSD Daemones required to report a Ceph OSD Daemon -``down`` to a Ceph Monitor by adding an ``mon osd min down reporters`` setting -(``osd min down reporters`` prior to v0.62) under the ``[mon]`` section of your -Ceph configuration file, or by setting the value at runtime. - - -.. ditaa:: +---------+ +---------+ - | OSD 1 | | Monitor | - +---------+ +---------+ - | | - | OSD 2 Is Down | - |-------------->| - | | - | OSD 2 Is Down | - |-------------->| - | | - | OSD 2 Is Down | - |-------------->| - | | - | |----------+ Mark - | | | OSD 2 - | |<---------+ Down +By default, two Ceph OSD Daemons from different hosts must report to the Ceph +Monitors that another Ceph OSD Daemon is ``down`` before the Ceph Monitors +acknowledge that the reported Ceph OSD Daemon is ``down``. But there is chance +that all the OSDs reporting the failure are hosted in a rack with a bad switch +which has trouble connecting to another OSD. To avoid this sort of false alarm, +we consider the peers reporting a failure a proxy for a potential "subcluster" +over the overall cluster that is similarly laggy. This is clearly not true in +all cases, but will sometimes help us localize the grace correction to a subset +of the system that is unhappy. ``mon osd reporter subtree level`` is used to +group the peers into the "subcluster" by their common ancestor type in CRUSH +map. By default, only two reports from different subtree are required to report +another Ceph OSD Daemon ``down``. You can change the number of reporters from +unique subtrees and the common ancestor type required to report a Ceph OSD +Daemon ``down`` to a Ceph Monitor by adding an ``mon osd min down reporters`` +and ``mon osd reporter subtree level`` settings under the ``[mon]`` section of +your Ceph configuration file, or by setting the value at runtime. + + +.. ditaa:: +---------+ +---------+ +---------+ + | OSD 1 | | OSD 2 | | Monitor | + +---------+ +---------+ +---------+ + | | | + | OSD 3 Is Down | | + |---------------+--------------->| + | | | + | | | + | | OSD 3 Is Down | + | |--------------->| + | | | + | | | + | | |---------+ Mark + | | | | OSD 3 + | | |<--------+ Down .. index:: peering failure @@ -118,13 +125,13 @@ setting the value at runtime. +---------+ +---------+ +-------+ +---------+ | | | | | Request To | | | - | Peer | | | + | Peer | | | |-------------->| | | |<--------------| | | | Peering | | | | | | Request To | | - | Peer | | + | Peer | | |----------------------------->| | | | |----+ OSD Monitor | @@ -135,7 +142,7 @@ setting the value at runtime. |-------------------------------------------->| |<--------------------------------------------| | Receive New Cluster Map | - + .. index:: OSD status @@ -149,10 +156,10 @@ event such as a failure, a change in placement group stats, a change in ``up_thru`` or when it boots within 5 seconds. You can change the Ceph OSD Daemon minimum report interval by adding an ``osd mon report interval min`` setting under the ``[osd]`` section of your Ceph configuration file, or by -setting the value at runtime. A Ceph OSD Daemon sends a report to a Ceph -Monitor every 120 seconds irrespective of whether any notable changes occur. -You can change the Ceph Monitor report interval by adding an ``osd mon report -interval max`` setting under the ``[osd]`` section of your Ceph configuration +setting the value at runtime. A Ceph OSD Daemon sends a report to a Ceph +Monitor every 120 seconds irrespective of whether any notable changes occur. +You can change the Ceph Monitor report interval by adding an ``osd mon report +interval max`` setting under the ``[osd]`` section of your Ceph configuration file, or by setting the value at runtime. @@ -207,18 +214,18 @@ Monitor Settings ``mon osd min up ratio`` -:Description: The minimum ratio of ``up`` Ceph OSD Daemons before Ceph will +:Description: The minimum ratio of ``up`` Ceph OSD Daemons before Ceph will mark Ceph OSD Daemons ``down``. - + :Type: Double :Default: ``.3`` ``mon osd min in ratio`` -:Description: The minimum ratio of ``in`` Ceph OSD Daemons before Ceph will +:Description: The minimum ratio of ``in`` Ceph OSD Daemons before Ceph will mark Ceph OSD Daemons ``out``. - + :Type: Double :Default: ``.75`` @@ -237,6 +244,15 @@ Monitor Settings :Default: ``0.3`` + +``mon osd laggy max interval`` +:Description: Maximum value of ``laggy_interval`` in laggy estimations (in seconds). + Monitor uses an adaptive approach to evaluate the ``laggy_interval`` of + a certain OSD. This value will be used to calculate the grace time for + that OSD. +:Type: Integer +:Default: 300 + ``mon osd adjust heartbeat grace`` :Description: If set to ``true``, Ceph will scale based on laggy estimations. @@ -251,38 +267,38 @@ Monitor Settings :Default: ``true`` -``mon osd auto mark in`` +``mon osd auto mark in`` -:Description: Ceph will mark any booting Ceph OSD Daemons as ``in`` +:Description: Ceph will mark any booting Ceph OSD Daemons as ``in`` the Ceph Storage Cluster. :Type: Boolean :Default: ``false`` -``mon osd auto mark auto out in`` +``mon osd auto mark auto out in`` -:Description: Ceph will mark booting Ceph OSD Daemons auto marked ``out`` +:Description: Ceph will mark booting Ceph OSD Daemons auto marked ``out`` of the Ceph Storage Cluster as ``in`` the cluster. - + :Type: Boolean -:Default: ``true`` +:Default: ``true`` -``mon osd auto mark new in`` +``mon osd auto mark new in`` -:Description: Ceph will mark booting new Ceph OSD Daemons as ``in`` the +:Description: Ceph will mark booting new Ceph OSD Daemons as ``in`` the Ceph Storage Cluster. - + :Type: Boolean -:Default: ``true`` +:Default: ``true`` -``mon osd down out interval`` +``mon osd down out interval`` :Description: The number of seconds Ceph waits before marking a Ceph OSD Daemon ``down`` and ``out`` if it doesn't respond. - + :Type: 32-bit Integer :Default: ``600`` @@ -298,21 +314,30 @@ Monitor Settings :Default: ``rack`` -``mon osd report timeout`` +``mon osd report timeout`` -:Description: The grace period in seconds before declaring +:Description: The grace period in seconds before declaring unresponsive Ceph OSD Daemons ``down``. :Type: 32-bit Integer :Default: ``900`` -``mon osd min down reporters`` +``mon osd min down reporters`` -:Description: The minimum number of Ceph OSD Daemons required to report a +:Description: The minimum number of Ceph OSD Daemons required to report a ``down`` Ceph OSD Daemon. :Type: 32-bit Integer -:Default: ``1`` +:Default: ``2`` + + +``mon osd reporter subtree level`` + +:Description: In which level of parent bucket the reporters are counted. The OSDs + send failure reports to monitor if they find its peer is not responsive. + And monitor mark the reported OSD out and then down after a grace period. +:Type: String +:Default: ``host`` .. index:: OSD hearbeat @@ -322,63 +347,61 @@ OSD Settings ``osd heartbeat address`` -:Description: An Ceph OSD Daemon's network address for heartbeats. +:Description: An Ceph OSD Daemon's network address for heartbeats. :Type: Address :Default: The host address. -``osd heartbeat interval`` +``osd heartbeat interval`` :Description: How often an Ceph OSD Daemon pings its peers (in seconds). :Type: 32-bit Integer :Default: ``6`` -``osd heartbeat grace`` +``osd heartbeat grace`` :Description: The elapsed time when a Ceph OSD Daemon hasn't shown a heartbeat that the Ceph Storage Cluster considers it ``down``. This setting has to be set in both the [mon] and [osd] or [global] section so that it is read by both the MON and OSD daemons. - :Type: 32-bit Integer :Default: ``20`` -``osd mon heartbeat interval`` +``osd mon heartbeat interval`` -:Description: How often the Ceph OSD Daemon pings a Ceph Monitor if it has no +:Description: How often the Ceph OSD Daemon pings a Ceph Monitor if it has no Ceph OSD Daemon peers. :Type: 32-bit Integer -:Default: ``30`` +:Default: ``30`` -``osd mon report interval max`` +``osd mon report interval max`` :Description: The maximum time in seconds that a Ceph OSD Daemon can wait before it must report to a Ceph Monitor. :Type: 32-bit Integer -:Default: ``120`` +:Default: ``120`` -``osd mon report interval min`` +``osd mon report interval min`` :Description: The minimum number of seconds a Ceph OSD Daemon may wait - from startup or another reportable event before reporting + from startup or another reportable event before reporting to a Ceph Monitor. :Type: 32-bit Integer :Default: ``5`` -:Valid Range: Should be less than ``osd mon report interval max`` +:Valid Range: Should be less than ``osd mon report interval max`` -``osd mon ack timeout`` +``osd mon ack timeout`` -:Description: The number of seconds to wait for a Ceph Monitor to acknowledge a +:Description: The number of seconds to wait for a Ceph Monitor to acknowledge a request for statistics. :Type: 32-bit Integer -:Default: ``30`` - +:Default: ``30`` diff --git a/doc/rados/configuration/pool-pg-config-ref.rst b/doc/rados/configuration/pool-pg-config-ref.rst index 9cb2c8f54ff9..3c871f1df85c 100644 --- a/doc/rados/configuration/pool-pg-config-ref.rst +++ b/doc/rados/configuration/pool-pg-config-ref.rst @@ -42,15 +42,95 @@ Ceph configuration file. :Type: 32-bit Integer :Default: ``300`` +``mon pg min inactive`` -``osd pg bits`` +:Description: Issue a ``HEALTH_ERR`` in cluster log if the number of PGs stay + inactive longer than ``mon_pg_stuck_threshold`` exceeds this + setting. A non-positive number means disabled, never go into ERR. +:Type: Integer +:Default: ``1`` + + +``mon pg warn min per osd`` + +:Description: Issue a ``HEALTH_WARN`` in cluster log if the average number + of PGs per (in) OSD is under this number. (a non-positive number + disables this) +:Type: Integer +:Default: ``30`` + + +``mon pg warn max per osd`` + +:Description: Issue a ``HEALTH_WARN`` in cluster log if the average number + of PGs per (in) OSD is above this number. (a non-positive number + disables this) +:Type: Integer +:Default: ``300`` + + +``mon pg warn min objects`` + +:Description: Do not warn if the total number of objects in cluster is below + this number +:Type: Integer +:Default: ``1000`` + + +``mon pg warn min pool objects`` + +:Description: Do not warn on pools whose object number is below this number +:Type: Integer +:Default: ``1000`` + + +``mon pg check down all threshold`` + +:Description: Threshold of down OSDs percentage after which we check all PGs + for stale ones. +:Type: Float +:Default: ``0.5`` + + +``mon pg warn max object skew`` + +:Description: Issue a ``HEALTH_WARN`` in cluster log if the average object number + of a certain pool is greater than ``mon pg warn max object skew`` times + the average object number of the whole pool. (a non-positive number + disables this) +:Type: Float +:Default: ``10`` + + +``mon delta reset interval`` + +:Description: Seconds of inactivity before we reset the pg delta to 0. We keep + track of the delta of the used space of each pool, so, for + example, it would be easier for us to understand the progress of + recovery or the performance of cache tier. But if the is no + activity reported for a certain pool, we just reset the history of + deltas of that pool. +:Type: Integer +:Default: ``10`` + + +``mon osd max op age`` + +:Description: Maximum op age before we get concerned (make it a power of 2). + A ``HEALTH_WARN`` will be issued if a request has been blocked longer + than this limit. +:Type: Float +:Default: ``32.0`` + + +``osd pg bits`` :Description: Placement group bits per Ceph OSD Daemon. :Type: 32-bit Integer :Default: ``6`` -``osd pgp bits`` +``osd pgp bits`` :Description: The number of bits per Ceph OSD Daemon for PGPs. :Type: 32-bit Integer