From 40f81036cf87b4d506c4de7e596f6f7e81ad97d1 Mon Sep 17 00:00:00 2001 From: Zac Dover Date: Sun, 2 Oct 2022 20:48:36 +1000 Subject: [PATCH] doc/glossary.rst: alphabetize glossary terms This commit (finally) alphabetizes the terms in the glossary. This is not a grammar-correcting or usage-correcting commit. Signed-off-by: Zac Dover --- doc/glossary.rst | 279 ++++++++++++++++++++++++----------------------- 1 file changed, 143 insertions(+), 136 deletions(-) diff --git a/doc/glossary.rst b/doc/glossary.rst index e473895762ea..7a96a45f967c 100644 --- a/doc/glossary.rst +++ b/doc/glossary.rst @@ -14,186 +14,193 @@ reflect either technical terms or legacy ways of referring to Ceph systems. .. glossary:: - Ceph Project - The aggregate term for the people, software, mission and infrastructure - of Ceph. - - cephx - The Ceph authentication protocol. Cephx operates like Kerberos, but it - has no single point of failure. + bluestore + OSD BlueStore is a new back end for OSD daemons (kraken and + newer versions). Unlike :term:`filestore` it stores objects + directly on the Ceph block devices without any file system + interface. Ceph - Ceph Platform - All Ceph software, which includes any piece of code hosted at - `https://github.com/ceph`_. - - Ceph System - Ceph Stack - A collection of two or more components of Ceph. + Ceph Block Device + Ceph Block Storage + The block storage "product," service or capabilities when used + in conjunction with ``librbd``, a hypervisor such as QEMU or + Xen, and a hypervisor abstraction layer such as ``libvirt``. - Ceph Node - Node - Host - Any single machine or server in a Ceph System. + Ceph Client + The collection of Ceph components which can access a Ceph + Storage Cluster. These include the Ceph Object Gateway, the + Ceph Block Device, the Ceph File System, and their + corresponding libraries, kernel modules, and FUSEs. - Ceph Storage Cluster - Ceph Object Store - RADOS - RADOS Cluster - Reliable Autonomic Distributed Object Store - The core set of storage software which stores the user's data (MON+OSD). + Ceph Client Libraries + The collection of libraries that can be used to interact with + components of the Ceph System. + Ceph Clients Ceph Cluster Map - Cluster Map - The set of maps comprising the monitor map, OSD map, PG map, MDS map and - CRUSH map. See `Cluster Map`_ for details. - - Ceph Object Storage - The object storage "product", service or capabilities, which consists - essentially of a Ceph Storage Cluster and a Ceph Object Gateway. - - Ceph Object Gateway - RADOS Gateway - RGW - The S3/Swift gateway component of Ceph. - - Ceph Block Device - RBD - The block storage component of Ceph. - - Ceph Block Storage - The block storage "product," service or capabilities when used in - conjunction with ``librbd``, a hypervisor such as QEMU or Xen, and a - hypervisor abstraction layer such as ``libvirt``. - + Ceph Dashboard Ceph File System CephFS Ceph FS - The POSIX filesystem components of Ceph. Refer - :ref:`CephFS Architecture ` and :ref:`ceph-file-system` for - more details. + The POSIX filesystem components of Ceph. Refer :ref:`CephFS + Architecture ` and :ref:`ceph-file-system` for + more details. - Cloud Platforms - Cloud Stacks - Third party cloud provisioning platforms such as OpenStack, CloudStack, - OpenNebula, Proxmox VE, etc. + Ceph Interim Release + Versions of Ceph that have not yet been put through quality + assurance testing, but may contain new features. - Object Storage Device - OSD - A physical or logical storage unit (*e.g.*, LUN). - Sometimes, Ceph users use the - term "OSD" to refer to :term:`Ceph OSD Daemon`, though the - proper term is "Ceph OSD". + Ceph Kernel Modules + The collection of kernel modules which can be used to interact + with the Ceph System (e.g., ``ceph.ko``, ``rbd.ko``). - Ceph OSD Daemon - Ceph OSD Daemons + Ceph Manager + Ceph Manager Dashboard + Ceph Metadata Server + Ceph Monitor + Ceph Node + Ceph Object Gateway + Ceph Object Storage + The object storage "product", service or capabilities, which + consists essentially of a Ceph Storage Cluster and a Ceph Object + Gateway. + + Ceph Object Store Ceph OSD The Ceph OSD software, which interacts with a logical disk (:term:`OSD`). Sometimes, Ceph users use the term "OSD" to refer to "Ceph OSD Daemon", though the proper term is "Ceph OSD". - OSD id - The integer that defines an OSD. It is generated by the monitors as part - of the creation of a new OSD. + Ceph OSD Daemon + Ceph OSD Daemons + Ceph Platform + All Ceph software, which includes any piece of code hosted at + `https://github.com/ceph`_. - OSD fsid - This is a unique identifier used to further improve the uniqueness of an - OSD and it is found in the OSD path in a file called ``osd_fsid``. This - ``fsid`` term is used interchangeably with ``uuid`` + Ceph Point Release + Any ad-hoc release that includes only bug or security fixes. - OSD uuid - Just like the OSD fsid, this is the OSD unique identifier and is used - interchangeably with ``fsid`` + Ceph Project + The aggregate term for the people, software, mission and + infrastructure of Ceph. - bluestore - OSD BlueStore is a new back end for OSD daemons (kraken and newer - versions). Unlike :term:`filestore` it stores objects directly on the - Ceph block devices without any file system interface. + Ceph Release + Any distinct numbered version of Ceph. - filestore - A back end for OSD daemons, where a Journal is needed and files are - written to the filesystem. + Ceph Release Candidate + A major version of Ceph that has undergone initial quality + assurance testing and is ready for beta testers. - Ceph Monitor - MON - The Ceph monitor software. + Ceph Stable Release + A major version of Ceph where all features from the preceding + interim releases have been put through quality assurance + testing successfully. - Ceph Manager - MGR - The Ceph manager software, which collects all the state from the whole - cluster in one place. + Ceph Stack + A collection of two or more components of Ceph. - Ceph Manager Dashboard - Ceph Dashboard - Dashboard Module - Dashboard Plugin - Dashboard - A built-in web-based Ceph management and monitoring application to - administer various aspects and objects of the cluster. The dashboard is - implemented as a Ceph Manager module. See :ref:`mgr-dashboard` for more - details. + Ceph Storage Cluster + Ceph System + Ceph Test Framework + cephx + The Ceph authentication protocol. Cephx operates like Kerberos, + but it has no single point of failure. - Ceph Metadata Server - MDS - The Ceph metadata software. + Cloud Platforms + Cloud Stacks + Third party cloud provisioning platforms such as OpenStack, + CloudStack, OpenNebula, Proxmox VE, etc. - Ceph Clients - Ceph Client - The collection of Ceph components which can access a Ceph Storage - Cluster. These include the Ceph Object Gateway, the Ceph Block Device, - the Ceph File System, and their corresponding libraries, kernel modules, - and FUSEs. + Cluster Map + The set of maps comprising the monitor map, OSD map, PG map, + MDS map and CRUSH map. See `Cluster Map`_ for details. - Ceph Kernel Modules - The collection of kernel modules which can be used to interact with the - Ceph System (e.g., ``ceph.ko``, ``rbd.ko``). + CRUSH + Controlled Replication Under Scalable Hashing. It is the + algorithm Ceph uses to compute object storage locations. - Ceph Client Libraries - The collection of libraries that can be used to interact with components - of the Ceph System. + CRUSH rule + The CRUSH data placement rule that applies to a particular + pool(s). - Ceph Release - Any distinct numbered version of Ceph. + Dashboard + A built-in web-based Ceph management and monitoring application + to administer various aspects and objects of the cluster. The + dashboard is implemented as a Ceph Manager module. See + :ref:`mgr-dashboard` for more details. - Ceph Point Release - Any ad-hoc release that includes only bug or security fixes. + Dashboard Module + Dashboard Plugin + filestore + A back end for OSD daemons, where a Journal is needed and files + are written to the filesystem. - Ceph Interim Release - Versions of Ceph that have not yet been put through quality assurance - testing, but may contain new features. + Host + Any single machine or server in a Ceph System. - Ceph Release Candidate - A major version of Ceph that has undergone initial quality assurance - testing and is ready for beta testers. + LVM tags + Extensible metadata for LVM volumes and groups. It is used to + store Ceph-specific information about devices and its + relationship with OSDs. - Ceph Stable Release - A major version of Ceph where all features from the preceding interim - releases have been put through quality assurance testing successfully. + MDS + The Ceph metadata software. - Ceph Test Framework - Teuthology - The collection of software that performs scripted tests on Ceph. + MGR + The Ceph manager software, which collects all the state from + the whole cluster in one place. - CRUSH - Controlled Replication Under Scalable Hashing. It is the algorithm - Ceph uses to compute object storage locations. + MON + The Ceph monitor software. - CRUSH rule - The CRUSH data placement rule that applies to a particular pool(s). + Node + Object Storage Device + OSD + A physical or logical storage unit (*e.g.*, LUN). + Sometimes, Ceph users use the + term "OSD" to refer to :term:`Ceph OSD Daemon`, though the + proper term is "Ceph OSD". + + OSD fsid + This is a unique identifier used to further improve the + uniqueness of an OSD and it is found in the OSD path in a file + called ``osd_fsid``. This ``fsid`` term is used interchangeably + with ``uuid`` + + OSD id + The integer that defines an OSD. It is generated by the + monitors as part of the creation of a new OSD. + + OSD uuid + Just like the OSD fsid, this is the OSD unique identifier and + is used interchangeably with ``fsid`` Pool Pools Pools are logical partitions for storing objects. + RADOS + RADOS Cluster + RADOS Gateway + RBD + The block storage component of Ceph. + + Reliable Autonomic Distributed Object Store + The core set of storage software which stores the user's data + (MON+OSD). + + RGW + The S3/Swift gateway component of Ceph. + systemd oneshot - A systemd ``type`` where a command is defined in ``ExecStart`` which will - exit upon completion (it is not intended to daemonize) + A systemd ``type`` where a command is defined in ``ExecStart`` + which will exit upon completion (it is not intended to + daemonize) - LVM tags - Extensible metadata for LVM volumes and groups. It is used to store - Ceph-specific information about devices and its relationship with - OSDs. + Teuthology + The collection of software that performs scripted tests on Ceph. .. _https://github.com/ceph: https://github.com/ceph .. _Cluster Map: ../architecture#cluster-map -- 2.47.3