diff options
Diffstat (limited to 'doc/glossary.rst')
-rw-r--r-- | doc/glossary.rst | 356 |
1 files changed, 356 insertions, 0 deletions
diff --git a/doc/glossary.rst b/doc/glossary.rst new file mode 100644 index 000000000..16594e6c2 --- /dev/null +++ b/doc/glossary.rst @@ -0,0 +1,356 @@ +=============== + Ceph Glossary +=============== + +.. glossary:: + + :ref:`BlueStore<rados_config_storage_devices_bluestore>` + OSD BlueStore is a storage back end used by OSD daemons, and + was designed specifically for use with Ceph. BlueStore was + introduced in the Ceph Kraken release. In the Ceph Luminous + release, BlueStore became Ceph's default storage back end, + supplanting FileStore. Unlike :term:`filestore`, BlueStore + stores objects directly on Ceph block devices without any file + system interface. Since Luminous (12.2), BlueStore has been + Ceph's default and recommended storage back end. + + Ceph + Ceph is a distributed network storage and file system with + distributed metadata management and POSIX semantics. + + Ceph Block Device + A software instrument that orchestrates the storage of + block-based data in Ceph. Ceph Block Device (also called "RBD", + or "RADOS block device") splits block-based application data + into "chunks". RADOS stores these chunks as objects. Ceph Block + Device orchestrates the storage of those objects across the + storage cluster. See also :term:`RBD`. + + Ceph Block Storage + One of the three kinds of storage supported by Ceph (the other + two are object storage and file storage). Ceph Block Storage is + the block storage "product", which refers to block-storage + related services and capabilities when used in conjunction with + the collection of (1) ``librbd`` (a python module that provides + file-like access to :term:`RBD` images), (2) a hypervisor such + as QEMU or Xen, and (3) a hypervisor abstraction layer such as + ``libvirt``. + + Ceph Client + Any of the Ceph components that can access a Ceph Storage + Cluster. This includes the Ceph Object Gateway, the Ceph Block + Device, the Ceph File System, and their corresponding + libraries. It also includes kernel modules, and FUSEs + (Filesystems in USERspace). + + Ceph Client Libraries + The collection of libraries that can be used to interact with + components of the Ceph Cluster. + + Ceph Cluster Map + See :term:`Cluster Map` + + Ceph Dashboard + :ref:`The Ceph Dashboard<mgr-dashboard>` is a built-in + web-based Ceph management and monitoring application through + which you can inspect and administer various resources within + the cluster. It is implemented as a :ref:`ceph-manager-daemon` + module. + + Ceph File System + See :term:`CephFS` + + :ref:`CephFS<ceph-file-system>` + The **Ceph F**\ile **S**\ystem, or CephFS, is a + POSIX-compliant file system built on top of Ceph’s distributed + object store, RADOS. See :ref:`CephFS Architecture + <arch-cephfs>` for more details. + + Ceph Interim Release + See :term:`Releases`. + + Ceph Kernel Modules + The collection of kernel modules that can be used to interact + with the Ceph Cluster (for example: ``ceph.ko``, ``rbd.ko``). + + :ref:`Ceph Manager<ceph-manager-daemon>` + The Ceph manager daemon (ceph-mgr) is a daemon that runs + alongside monitor daemons to provide monitoring and interfacing + to external monitoring and management systems. Since the + Luminous release (12.x), no Ceph cluster functions properly + unless it contains a running ceph-mgr daemon. + + Ceph Manager Dashboard + See :term:`Ceph Dashboard`. + + Ceph Metadata Server + See :term:`MDS`. + + Ceph Monitor + A daemon that maintains a map of the state of the cluster. This + "cluster state" includes the monitor map, the manager map, the + OSD map, and the CRUSH map. A Ceph cluster must contain a + minimum of three running monitors in order to be both redundant + and highly-available. Ceph monitors and the nodes on which they + run are often referred to as "mon"s. See :ref:`Monitor Config + Reference <monitor-config-reference>`. + + Ceph Node + A Ceph node is a unit of the Ceph Cluster that communicates with + other nodes in the Ceph Cluster in order to replicate and + redistribute data. All of the nodes together are called the + :term:`Ceph Storage Cluster`. Ceph nodes include :term:`OSD`\s, + :term:`Ceph Monitor`\s, :term:`Ceph Manager`\s, and + :term:`MDS`\es. The term "node" is usually equivalent to "host" + in the Ceph documentation. If you have a running Ceph Cluster, + you can list all of the nodes in it by running the command + ``ceph node ls all``. + + :ref:`Ceph Object Gateway<object-gateway>` + An object storage interface built on top of librados. Ceph + Object Gateway provides a RESTful gateway between applications + and Ceph storage clusters. + + Ceph Object Storage + See :term:`Ceph Object Store`. + + Ceph Object Store + A Ceph Object Store consists of a :term:`Ceph Storage Cluster` + and a :term:`Ceph Object Gateway` (RGW). + + :ref:`Ceph OSD<rados_configuration_storage-devices_ceph_osd>` + Ceph **O**\bject **S**\torage **D**\aemon. The Ceph OSD + software, which interacts with logical disks (:term:`OSD`). + Around 2013, there was an attempt by "research and industry" + (Sage's own words) to insist on using the term "OSD" to mean + only "Object Storage Device", but the Ceph community has always + persisted in using the term to mean "Object Storage Daemon" and + no less an authority than Sage Weil himself confirms in + November of 2022 that "Daemon is more accurate for how Ceph is + built" (private correspondence between Zac Dover and Sage Weil, + 07 Nov 2022). + + Ceph OSD Daemon + See :term:`Ceph OSD`. + + Ceph OSD Daemons + See :term:`Ceph OSD`. + + Ceph Platform + All Ceph software, which includes any piece of code hosted at + `https://github.com/ceph`_. + + Ceph Point Release + See :term:`Releases`. + + Ceph Project + The aggregate term for the people, software, mission and + infrastructure of Ceph. + + Ceph Release + See :term:`Releases`. + + Ceph Release Candidate + See :term:`Releases`. + + Ceph Stable Release + See :term:`Releases`. + + Ceph Stack + A collection of two or more components of Ceph. + + :ref:`Ceph Storage Cluster<arch-ceph-storage-cluster>` + The collection of :term:`Ceph Monitor`\s, :term:`Ceph + Manager`\s, :term:`Ceph Metadata Server`\s, and :term:`OSD`\s + that work together to store and replicate data for use by + applications, Ceph Users, and :term:`Ceph Client`\s. Ceph + Storage Clusters receive data from :term:`Ceph Client`\s. + + cephx + The Ceph authentication protocol. Cephx operates like Kerberos, + but it has no single point of failure. + + Cloud Platforms + Cloud Stacks + Third party cloud provisioning platforms such as OpenStack, + CloudStack, OpenNebula, and Proxmox VE. + + Cluster Map + The set of maps consisting of the monitor map, OSD map, PG map, + MDS map, and CRUSH map, which together report the state of the + Ceph cluster. See :ref:`the "Cluster Map" section of the + Architecture document<architecture_cluster_map>` for details. + + CRUSH + Controlled Replication Under Scalable Hashing. It is the + algorithm Ceph uses to compute object storage locations. + + CRUSH rule + The CRUSH data placement rule that applies to a particular + pool(s). + + DAS + **D**\irect-\ **A**\ttached **S**\torage. Storage that is + attached directly to the computer accessing it, without passing + through a network. Contrast with NAS and SAN. + + :ref:`Dashboard<mgr-dashboard>` + A built-in web-based Ceph management and monitoring application + to administer various aspects and objects of the cluster. The + dashboard is implemented as a Ceph Manager module. See + :ref:`mgr-dashboard` for more details. + + Dashboard Module + Another name for :term:`Dashboard`. + + Dashboard Plugin + filestore + A back end for OSD daemons, where a Journal is needed and files + are written to the filesystem. + + FQDN + **F**\ully **Q**\ualified **D**\omain **N**\ame. A domain name + that is applied to a node in a network and that specifies the + node's exact location in the tree hierarchy of the DNS. + + In the context of Ceph cluster administration, FQDNs are often + applied to hosts. In this documentation, the term "FQDN" is + used mostly to distinguish between FQDNs and relatively simpler + hostnames, which do not specify the exact location of the host + in the tree hierarchy of the DNS but merely name the host. + + Host + Any single machine or server in a Ceph Cluster. See :term:`Ceph + Node`. + + LVM tags + Extensible metadata for LVM volumes and groups. It is used to + store Ceph-specific information about devices and its + relationship with OSDs. + + :ref:`MDS<cephfs_add_remote_mds>` + The Ceph **M**\eta\ **D**\ata **S**\erver daemon. Also referred + to as "ceph-mds". The Ceph metadata server daemon must be + running in any Ceph cluster that runs the CephFS file system. + The MDS stores all filesystem metadata. + + MGR + The Ceph manager software, which collects all the state from + the whole cluster in one place. + + MON + The Ceph monitor software. + + Node + See :term:`Ceph Node`. + + Object Storage Device + See :term:`OSD`. + + OSD + Probably :term:`Ceph OSD`, but not necessarily. Sometimes + (especially in older correspondence, and especially in + documentation that is not written specifically for Ceph), "OSD" + means "**O**\bject **S**\torage **D**\evice", which refers to a + physical or logical storage unit (for example: LUN). The Ceph + community has always used the term "OSD" to refer to + :term:`Ceph OSD Daemon` despite an industry push in the + mid-2010s to insist that "OSD" should refer to "Object Storage + Device", so it is important to know which meaning is intended. + + OSD fsid + This is a unique identifier used to identify an OSD. It is + found in the OSD path in a file called ``osd_fsid``. The + term ``fsid`` is used interchangeably with ``uuid`` + + OSD id + The integer that defines an OSD. It is generated by the + monitors during the creation of each OSD. + + OSD uuid + This is the unique identifier of an OSD. This term is used + interchangeably with ``fsid`` + + :ref:`Pool<rados_pools>` + A pool is a logical partition used to store objects. + + Pools + See :term:`pool`. + + RADOS + **R**\eliable **A**\utonomic **D**\istributed **O**\bject + **S**\tore. RADOS is the object store that provides a scalable + service for variably-sized objects. The RADOS object store is + the core component of a Ceph cluster. `This blog post from + 2009 + <https://ceph.io/en/news/blog/2009/the-rados-distributed-object-store/>`_ + provides a beginner's introduction to RADOS. Readers interested + in a deeper understanding of RADOS are directed to `RADOS: A + Scalable, Reliable Storage Service for Petabyte-scale Storage + Clusters <https://ceph.io/assets/pdfs/weil-rados-pdsw07.pdf>`_. + + RADOS Cluster + A proper subset of the Ceph Cluster consisting of + :term:`OSD`\s, :term:`Ceph Monitor`\s, and :term:`Ceph + Manager`\s. + + RADOS Gateway + See :term:`RGW`. + + RBD + The block storage component of Ceph. Also called "RADOS Block + Device" or :term:`Ceph Block Device`. + + Releases + + Ceph Interim Release + A version of Ceph that has not yet been put through + quality assurance testing. May contain new features. + + Ceph Point Release + Any ad hoc release that includes only bug fixes and + security fixes. + + Ceph Release + Any distinct numbered version of Ceph. + + Ceph Release Candidate + A major version of Ceph that has undergone initial + quality assurance testing and is ready for beta + testers. + + Ceph Stable Release + A major version of Ceph where all features from the + preceding interim releases have been put through + quality assurance testing successfully. + + Reliable Autonomic Distributed Object Store + The core set of storage software which stores the user's data + (MON+OSD). See also :term:`RADOS`. + + :ref:`RGW<object-gateway>` + **R**\ADOS **G**\ate **W**\ay. + + The component of Ceph that provides a gateway to both the + Amazon S3 RESTful API and the OpenStack Swift API. Also called + "RADOS Gateway" and "Ceph Object Gateway". + + secrets + Secrets are credentials used to perform digital authentication + whenever privileged users must access systems that require + authentication. Secrets can be passwords, API keys, tokens, SSH + keys, private certificates, or encryption keys. + + SDS + Software-defined storage. + + systemd oneshot + A systemd ``type`` where a command is defined in ``ExecStart`` + which will exit upon completion (it is not intended to + daemonize) + + Teuthology + The collection of software that performs scripted tests on Ceph. + +.. _https://github.com/ceph: https://github.com/ceph +.. _Cluster Map: ../architecture#cluster-map |