From: Nilamdyuti Goswami Date: Fri, 12 Dec 2014 20:54:41 +0000 (+0530) Subject: doc: Adds man page for ceph. X-Git-Tag: v0.80.9~6^2~2 X-Git-Url: http://git-server-git.apps.pok.os.sepia.ceph.com/?a=commitdiff_plain;h=893e5cd1f4fd5ea77d65f0e78cc8cff88eb19c1a;p=ceph.git doc: Adds man page for ceph. Signed-off-by: Nilamdyuti Goswami (cherry picked from commit 76da87a64ca6b3cc0ceeaf63e19a9f440d6f4161) --- diff --git a/doc/man/8/ceph.rst b/doc/man/8/ceph.rst index 65f4bed0001..2b52df3e681 100644 --- a/doc/man/8/ceph.rst +++ b/doc/man/8/ceph.rst @@ -1,42 +1,716 @@ -========================================== - ceph -- ceph file system control utility -========================================== +================================== + ceph -- ceph administration tool +================================== .. program:: ceph Synopsis ======== -| **ceph** [ -m *monaddr* ] [ -w | *command* ... ] +| **ceph** **auth** *add* ** {** [**...]} +| **ceph** **auth** *export* ** + +| **ceph** **config-key** *get* ** + +| **ceph** **mds** *add_data_pool* ** + +| **ceph** **mds** *getmap* {**} + +| **ceph** **mon** *add* ** <*IPaddr[:port]*> + +| **ceph** **mon_status** + +| **ceph** **osd** *create* {**} + +| **ceph** **osd** **crush** *add* ** +** ** [**...] + +| **ceph** **pg** *force_create_pg* ** + +| **ceph** **pg** *stat* + +| **ceph** **quorum_status** Description =========== -**ceph** is a control utility for communicating with the monitor -cluster of a running Ceph distributed storage system. +**ceph** is a control utility which is used for manual deployment and maintenance +of a Ceph cluster. It provides a diverse set of commands that allows deployment of +monitors, OSDs, placement groups, MDS and overall maintenance, administration +of the cluster. + +Commands +======== + +**auth**: Manage authentication keys. It is used for adding, removing, exporting +or updating of authentication keys for a particular entity such as a monitor or +OSD. It uses some additional subcommands. + +Subcommand **add** adds authentication info for a particular entity from input +file, or random key if no input given and/or any caps specified in the command. + +Usage: ceph auth add { [...]} + +Subcommand **caps** updates caps for **name** from caps specified in the command. + +Usage: ceph auth caps [...] + +Subcommand **del** deletes all caps for **name**. + +Usage: ceph auth del + +Subcommand **export** writes keyring for requested entity, or master keyring if +none given. + +Usage: ceph auth export {} + +Subcommand **get** writes keyring file with requested key. + +Usage: ceph auth get + +Subcommand **get-key** displays requested key. + +Usage: ceph auth get-key + +Subcommand **get-or-create** adds authentication info for a particular entity +from input file, or random key if no input given and/or any caps specified in the +command. + +Usage: ceph auth get-or-create { [...]} + +Subcommand **get-or-create-key** gets or adds key for **name** from system/caps +pairs specified in the command. If key already exists, any given caps must match +the existing caps for that key. + +Subcommand **import** reads keyring from input file. + +Usage: ceph auth import + +Subcommand **list** lists authentication state. + +Usage: ceph auth list + +Subcommand **print-key** displays requested key. + +Usage: ceph auth print-key + +Subcommand **print_key** displays requested key. + +Usage: ceph auth print_key + +**compact**: Causes compaction of monitor's leveldb storage. + +Usage: ceph compact + +**config-key**: Manage configuration key. It uses some additional subcommands. + +Subcommand **get** gets the configuration key. + +Usage: ceph config-key get + +Subcommand **put** puts configuration key and values. + +Usage: ceph config-key put {} + +Subcommand **exists** checks for configuration keys existence. + +Usage: ceph config-key exists + +Subcommand **list** lists configuration keys. + +Usage: ceph config-key list + +Subcommand **del** deletes configuration key. + +Usage: ceph config-key del + +**df**: Show cluster's free space status. + +Usage: ceph df + +**fsid**: Show cluster's FSID/UUID. + +Usage: ceph fsid + +**health**: Show cluster's health. + +Usage: ceph health + +**heap**: Show heap usage info (available only if compiled with tcmalloc) + +Usage: ceph heap dump|start_profiler|stop_profiler|release|stats + +**injectargs**: Inject configuration arguments into monitor. + +Usage: ceph injectargs [...] + +**log**: Log supplied text to the monitor log. + +Usage: ceph log [...] + +**mds**: Manage metadata server configuration and administration. It uses some +additional subcommands. + +Subcommand **add_data_pool** adds data pool. + +Usage: ceph mds add_data_pool + +Subcommand **cluster_down** takes mds cluster down. + +Usage: ceph mds cluster_down + +Subcommand **cluster_up** brings mds cluster up. + +Usage: ceph mds cluster_up + +Subcommand **compat** manages compatible features. It uses some additional +subcommands. + +Subcommand **rm_compat** removes compatible feature. + +Usage: ceph mds compat rm_compat + +Subcommand **rm_incompat** removes incompatible feature. + +Usage: ceph mds compat rm_incompat + +Subcommand **show** shows mds compatibility settings. + +Usage: ceph mds compat show + +Subcommand **deactivate** stops mds. + +Usage: ceph mds deactivate + +Subcommand **dump** dumps information, optionally from epoch. + +Usage: ceph mds dump {} + +Subcommand **fail** forces mds to status fail. + +Usage: ceph mds fail + +Subcommand **getmap** gets MDS map, optionally from epoch. + +Usage: ceph mds getmap {} + +Subcommand **newfs** makes new filesystem using pools and . + +Usage: ceph mds newfs {--yes-i-really-mean-it} + +Subcommand **remove_data_pool** removes data pool. + +Usage: ceph mds remove_data_pool + +Subcommand **rm** removes inactive mds. + +Usage: ceph mds rm (type.id)> + +Subcommand **rmfailed** removes failed mds. + +Usage: ceph mds rmfailed + +Subcommand **set_max_mds** sets max MDS index. + +Usage: ceph mds set_max_mds + +Subcommand **set_state** sets mds state of to . + +Usage: ceph mds set_state + +Subcommand **setmap** sets mds map; must supply correct epoch number. + +Usage: ceph mds setmap + +Subcommand **stat** shows MDS status. + +Usage: ceph mds stat + +Subcommand **stop** stops mds. + +Usage: ceph mds stop + +Subcommand **tell** sends command to particular mds. + +Usage: ceph mds tell [...] + +**mon**: Manage monitor configuration and administration. It uses some +additional subcommands. + +Subcommand **add** adds new monitor named at . + +Usage: ceph mon add + +Subcommand **dump** dumps formatted monmap (optionally from epoch) + +Usage: ceph mon dump {} + +Subcommand **getmap** gets monmap. + +Usage: ceph mon getmap {} + +Subcommand **remove** removes monitor named . + +Usage: ceph mon remove + +Subcommand **stat** summarizes monitor status. + +Usage: ceph mon stat + +Subcommand **mon_status** reports status of monitors. + +Usage: ceph mon_status + +**osd**: Manage OSD configuration and administration. It uses some additional +subcommands. + +Subcommand **create** creates new osd (with optional UUID). + +Usage: ceph osd create {} + +Subcommand **crush** is used for CRUSH management. It uses some additional +subcommands. + +Subcommand **add** adds or updates crushmap position and weight for with + and location . + +Usage: ceph osd crush add [...] + +Subcommand **add-bucket** adds no-parent (probably root) crush bucket of +type . + +Usage: ceph osd crush add-bucket + +Subcommand **create-or-move** creates entry or moves existing entry for + at/to location . + +Usage: ceph osd crush create-or-move +[...] + +Subcommand **dump** dumps crush map. + +Usage: ceph osd crush dump + +Subcommand **link** links existing entry for under location . + +Usage: ceph osd crush link [...] + +Subcommand **move** moves existing entry for to location . + +Usage: ceph osd crush move [...] + +Subcommand **remove** removes from crush map (everywhere, or just at +). + +Usage: ceph osd crush remove {} + +Subcommand **reweight** change 's weight to in crush map. + +Usage: ceph osd crush reweight + +Subcommand **rm** removes from crush map (everywhere, or just at +). + +Usage: ceph osd crush rm {} + +Subcommand **rule** is used for creating crush rules. It uses some additional +subcommands. + +Subcommand **create-erasure** creates crush rule for erasure coded pool +created with (default default). + +Usage: ceph osd crush rule create-erasure {} + +Subcommand **create-simple** creates crush rule to start from , +replicate across buckets of type , using a choose mode of +(default firstn; indep best for erasure pools). + +Usage: ceph osd crush rule create-simple {firstn|indep} + +Subcommand **dump** dumps crush rule (default all). + +Usage: ceph osd crush rule dump {} + +Subcommand **list** lists crush rules. + +Usage: ceph osd crush rule list + +Subcommand **ls** lists crush rules. + +Usage: ceph osd crush rule ls + +Subcommand **rm** removes crush rule . + +Usage: ceph osd crush rule rm + +Subcommand **set** sets crush map from input file. + +Usage: ceph osd crush set + +Subcommand **set** with osdname/osd.id update crushmap position and weight +for to with location . + +Usage: ceph osd crush set [...] + +Subcommand **show-tunables** shows current crush tunables. + +Usage: ceph osd crush show-tunables + +Subcommand **tunables** sets crush tunables values to . + +Usage: ceph osd crush tunables legacy|argonaut|bobtail|firefly|optimal|default + +Subcommand **unlink** unlinks from crush map (everywhere, or just at +). + +Usage: ceph osd crush unlink {} + +Subcommand **deep-scrub** initiates deep scrub on specified osd. + +Usage: ceph osd deep-scrub + +Subcommand **down** sets osd(s) [...] down. + +Usage: ceph osd down [...] + +Subcommand **dump** prints summary of OSD map. + +Usage: ceph osd dump {} + +Subcommand **erasure-code-profile** is used for managing the erasure code +profiles. It uses some additional subcommands. -There are three basic modes of operation. +Subcommand **get** gets erasure code profile . -Interactive mode ----------------- +Usage: ceph osd erasure-code-profile get -To start in interactive mode, no arguments are necessary. Control-d or -'quit' will exit. +Subcommand **ls** lists all erasure code profiles. -Watch mode ----------- +Usage: ceph osd erasure-code-profile ls -Watch mode shows cluster state changes as they occur. For example:: +Subcommand **rm** removes erasure code profile . - ceph -w +Usage: ceph osd erasure-code-profile rm -Command line mode ------------------ +Subcommand **set** creates erasure code profile with [ ...] +pairs. Add a --force at the end to override an existing profile (IT IS RISKY). -Finally, to send a single instruction to the monitor cluster (and wait -for a response), the command can be specified on the command line. +Usage: ceph osd erasure-code-profile set { [...]} +Subcommand **find** find osd in the CRUSH map and shows its location. + +Usage: ceph osd find + +Subcommand **getcrushmap** gets CRUSH map. + +Usage: ceph osd getcrushmap {} + +Subcommand **getmap** gets OSD map. + +Usage: ceph osd getmap {} + +Subcommand **getmaxosd** shows largest OSD id. + +Usage: ceph osd getmaxosd + +Subcommand **in** sets osd(s) [...] in. + +Usage: ceph osd in [...] + +Subcommand **lost** marks osd as permanently lost. THIS DESTROYS DATA IF NO +MORE REPLICAS EXIST, BE CAREFUL. + +Usage: ceph osd lost {--yes-i-really-mean-it} + +Subcommand **ls** shows all OSD ids. + +Usage: ceph osd ls {} + +Subcommand **lspools** lists pools. + +Usage: ceph osd lspools {} + +Subcommand **map** finds pg for in . + +Usage: ceph osd map + +Subcommand **metadata** fetches metadata for osd . + +Usage: ceph osd metadata + +Subcommand **out** sets osd(s) [...] out. + +Usage: ceph osd out [...] + +Subcommand **pause** pauses osd. + +Usage: ceph osd pause + +Subcommand **perf** prints dump of OSD perf summary stats. + +Usage: ceph osd perf + +Subcommand **pg-temp** set pg_temp mapping pgid:[ [...]] (developers +only). + +Usage: ceph osd pg-temp { [...]} + +Subcommand **pool** is used for managing data pools. It uses some additional +subcommands. + +Subcommand **create** creates pool. + +Usage: ceph osd pool create {} {replicated|erasure} +{} {} + +Subcommand **delete** deletes pool. + +Usage: ceph osd pool delete {} {--yes-i-really-really-mean-it} + +Subcommand **get** gets pool parameter . + +Usage: ceph osd pool get size|min_size|crash_replay_interval|pg_num| +pgp_num|crush_ruleset|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp| + +ceph osd pool get auid|target_max_objects|target_max_bytes + +ceph osd pool get cache_target_dirty_ratio|cache_target_full_ratio + +ceph osd pool get cache_min_flush_age|cache_min_evict_age| +erasure_code_profile + +Subcommand **get-quota** obtains object or byte limits for pool. + +Usage: ceph osd pool get-quota + +Subcommand **mksnap** makes snapshot in . + +Usage: ceph osd pool mksnap + +Subcommand **rename** renames to . + +Usage: ceph osd pool rename + +Subcommand **rmsnap** removes snapshot from . + +Usage: ceph osd pool rmsnap + +Subcommand **set** sets pool parameter to . + +Usage: ceph osd pool set size|min_size|crash_replay_interval|pg_num| +pgp_num|crush_ruleset|hashpspool|hit_set_type|hit_set_period| + +ceph osd pool set hit_set_count|hit_set_fpp|debug_fake_ec_pool + +ceph osd pool set target_max_bytes|target_max_objects + +ceph osd pool set cache_target_dirty_ratio|cache_target_full_ratio + +ceph osd pool set cache_min_flush_age + +ceph osd pool set cache_min_evict_age|auid {--yes-i-really-mean-it} + +Subcommand **set-quota** sets object or byte limit on pool. + +Usage: ceph osd pool set-quota max_objects|max_bytes + +Subcommand **stats** obtain stats from all pools, or from specified pool. + +Usage: ceph osd pool stats {} + +Subcommand **primary-affinity** adjust osd primary-affinity from 0.0 <= +<= 1.0 + +Usage: ceph osd primary-affinity + +Subcommand **primary-temp** sets primary_temp mapping pgid:|-1 (developers +only). + +Usage: ceph osd primary-temp + +Subcommand **repair** initiates repair on a specified osd. + +Usage: ceph osd repair + +Subcommand **reweight** reweights osd to 0.0 < < 1.0. + +Usage: osd reweight + +Subcommand **reweight-by-utilization** reweight OSDs by utilization +[overload-percentage-for-consideration, default 120]. + +Usage: ceph osd reweight-by-utilization {} + +Subcommand **rm** removes osd(s) [...] in the cluster. + +Usage: ceph osd rm [...] + +Subcommand **scrub** initiates scrub on specified osd. + +Usage: ceph osd scrub + +Subcommand **set** sets . + +Usage: ceph osd set pause|noup|nodown|noout|noin|nobackfill|norecover|noscrub| +nodeep-scrub|notieragent + +Subcommand **setcrushmap** sets crush map from input file. + +Usage: ceph osd setcrushmap + +Subcommand **setmaxosd** sets new maximum osd value. + +Usage: ceph osd setmaxosd + +Subcommand **stat** prints summary of OSD map. + +Usage: ceph osd stat + +Subcommand **thrash** thrashes OSDs for . + +Usage: ceph osd thrash + +Subcommand **tier** is used for managing tiers. It uses some additional +subcommands. + +Subcommand **add** adds the tier (the second one) to base pool +(the first one). + +Usage: ceph osd tier add {--force-nonempty} + +Subcommand **add-cache** adds a cache (the second one) of size +to existing pool (the first one). + +Usage: ceph osd tier add-cache + +Subcommand **cache-mode** specifies the caching mode for cache tier . + +Usage: ceph osd tier cache-mode none|writeback|forward|readonly + +Subcommand **remove** removes the tier (the second one) from base pool + (the first one). + +Usage: ceph osd tier remove + +Subcommand **remove-overlay** removes the overlay pool for base pool . + +Usage: ceph osd tier remove-overlay + +Subcommand **set-overlay** set the overlay pool for base pool to be +. + +Usage: ceph osd tier set-overlay + +Subcommand **tree** prints OSD tree. + +Usage: ceph osd tree {} + +Subcommand **unpause** unpauses osd. + +Usage: ceph osd unpause + +Subcommand **unset** unsets . + +Usage: osd unset pause|noup|nodown|noout|noin|nobackfill|norecover|noscrub| +nodeep-scrub|notieragent + +**pg**: It is used for managing the placement groups in OSDs. It uses some +additional subcommands. + +Subcommand **debug** shows debug info about pgs. + +Usage: ceph pg debug unfound_objects_exist|degraded_pgs_exist + +Subcommand **deep-scrub** starts deep-scrub on . + +Usage: ceph pg deep-scrub + +Subcommand **dump** shows human-readable versions of pg map (only 'all' valid with +plain). + +Usage: ceph pg dump {all|summary|sum|delta|pools|osds|pgs|pgs_brief} + +ceph pg dump {all|summary|sum|delta|pools|osds|pgs|pgs_brief...} + +Subcommand **dump_json** shows human-readable version of pg map in json only. + +Usage: ceph pg dump_json {all|summary|sum|pools|osds|pgs[all|summary|sum|pools| +osds|pgs...]} + +Subcommand **dump_pools_json** shows pg pools info in json only. + +Usage: ceph pg dump_pools_json + +Subcommand **dump_stuck** shows information about stuck pgs. + +Usage: ceph pg dump_stuck {inactive|unclean|stale[inactive|unclean|stale...]} +{} + +Subcommand **force_create_pg** forces creation of pg . + +Usage: ceph pg force_create_pg + +Subcommand **getmap** gets binary pg map to -o/stdout. + +Usage: ceph pg getmap + +Subcommand **map** shows mapping of pg to osds. + +Usage: ceph pg map + +Subcommand **repair** starts repair on . + +Usage: ceph pg repair + +Subcommand **scrub** starts scrub on . + +Usage: ceph pg scrub + +Subcommand **send_pg_creates** triggers pg creates to be issued. + +Usage: ceph pg send_pg_creates + +Subcommand **set_full_ratio** sets ratio at which pgs are considered full. + +Usage: ceph pg set_full_ratio + +Subcommand **set_nearfull_ratio** sets ratio at which pgs are considered nearly +full. + +Usage: ceph pg set_nearfull_ratio + +Subcommand **stat** shows placement group status. + +Usage: ceph pg stat + +**quorum**: Enter or exit quorum. + +Usage: ceph quorum enter|exit + +**quorum_status**: Reports status of monitor quorum. + +Usage: ceph quorum_status + +**report**: Reports full status of cluster, optional title tag strings. + +Usage: ceph report { [...]} + +**scrub**: Scrubs the monitor stores. + +Usage: ceph scrub + +**status**: Shows cluster status. + +Usage: ceph status + +**sync force**: Forces sync of and clear monitor store. + +Usage: ceph sync force {--yes-i-really-mean-it} {--i-know-what-i-am-doing} + +**tell**: Sends a command to a specific daemon. + +Usage: ceph tell [...] Options ======= @@ -58,30 +732,73 @@ Options Use ceph.conf configuration file instead of the default /etc/ceph/ceph.conf to determine monitor addresses during startup. -.. option:: -m monaddress[:port] +.. option:: --id CLIENT_ID, --user CLIENT_ID - Connect to specified monitor (instead of looking through ceph.conf). + Client id for authentication. +.. option:: --name CLIENT_NAME, -n CLIENT_NAME -Examples -======== + Client name for authentication. + +.. option:: --cluster CLUSTER + + Name of the Ceph cluster. + +.. option:: --admin-daemon ADMIN_SOCKET + + Submit admin-socket commands. + +.. option:: --admin-socket ADMIN_SOCKET_NOPE + + You probably mean --admin-daemon + +.. option:: -s, --status + + Show cluster status. + +.. option:: -w, --watch + + Watch live cluster changes. + +.. option:: --watch-debug + + Watch debug events. + +.. option:: --watch-info + + Watch info events. + +.. option:: --watch-sec + + Watch security events. + +.. option:: --watch-warn + + Watch warning events. + +.. option:: --watch-error + + Watch error events. + +.. option:: --version, -v + + Display version. -To grab a copy of the current OSD map:: +.. option:: --verbose - ceph -m 1.2.3.4:6789 osd getmap -o osdmap + Make verbose. -To get a dump of placement group (PG) state:: +.. option:: --concise - ceph pg dump -o pg.txt + Make less verbose. +.. option:: -f {json,json-pretty,xml,xml-pretty,plain}, --format -Monitor commands -================ + Format of output. -A more complete summary of commands understood by the monitor cluster can be found in the -online documentation, at +.. option:: --connect-timeout CLUSTER_TIMEOUT - http://ceph.com/docs/master/rados/operations/control + Set a timeout for connecting to the cluster. Availability @@ -94,4 +811,6 @@ http://ceph.com/docs for more information. See also ======== -:doc:`ceph `\(8), +:doc:`ceph-mon `\(8), +:doc:`ceph-osd `\(8), +:doc:`ceph-mds `\(8)