summaryrefslogtreecommitdiffstats
path: root/doc/cephfs/mds-journaling.rst
diff options
context:
space:
mode:
Diffstat (limited to 'doc/cephfs/mds-journaling.rst')
-rw-r--r--doc/cephfs/mds-journaling.rst158
1 files changed, 158 insertions, 0 deletions
diff --git a/doc/cephfs/mds-journaling.rst b/doc/cephfs/mds-journaling.rst
new file mode 100644
index 000000000..b6ccf27c8
--- /dev/null
+++ b/doc/cephfs/mds-journaling.rst
@@ -0,0 +1,158 @@
+MDS Journaling
+==============
+
+CephFS Metadata Pool
+--------------------
+
+CephFS uses a separate (metadata) pool for managing file metadata (inodes and
+dentries) in a Ceph File System. The metadata pool has all the information about
+files in a Ceph File System including the File System hierarchy. Additionally,
+CephFS maintains meta information related to other entities in a file system
+such as file system journals, open file table, session map, etc.
+
+This document describes how Ceph Metadata Servers use and rely on journaling.
+
+CephFS MDS Journaling
+---------------------
+
+CephFS metadata servers stream a journal of metadata events into RADOS in the metadata
+pool prior to executing a file system operation. Active MDS daemon(s) manage metadata
+for files and directories in CephFS.
+
+CephFS uses journaling for couple of reasons:
+
+#. Consistency: On an MDS failover, the journal events can be replayed to reach a
+ consistent file system state. Also, metadata operations that require multiple
+ updates to the backing store need to be journaled for crash consistency (along
+ with other consistency mechanisms such as locking, etc..).
+
+#. Performance: Journal updates are (mostly) sequential, hence updates to journals
+ are fast. Furthermore, updates can be batched into single write, thereby saving
+ disk seek time involved in updates to different parts of a file. Having a large
+ journal also helps a standby MDS to warm its cache which helps indirectly during
+ MDS failover.
+
+Each active metadata server maintains its own journal in the metadata pool. Journals
+are striped over multiple objects. Journal entries which are not required (deemed as
+old) are trimmed by the metadata server.
+
+Journal Events
+--------------
+
+Apart from journaling file system metadata updates, CephFS journals various other events
+such as client session info and directory import/export state to name a few. These events
+are used by the metadata sever to reestablish correct state as required, e.g., Ceph MDS
+tries to reconnect clients on restart when journal events get replayed and a specific
+event type in the journal specifies that a client entity type has a session with the MDS
+before it was restarted.
+
+To examine the list of such events recorded in the journal, CephFS provides a command
+line utility `cephfs-journal-tool` which can be used as follows:
+
+::
+
+ cephfs-journal-tool --rank=<fs>:<rank> event get list
+
+`cephfs-journal-tool` is also used to discover and repair a damaged Ceph File System.
+(See :doc:`/cephfs/cephfs-journal-tool` for more details)
+
+Journal Event Types
+-------------------
+
+Following are various event types that are journaled by the MDS.
+
+#. `EVENT_COMMITTED`: Mark a request (id) as committed.
+
+#. `EVENT_EXPORT`: Maps directories to an MDS rank.
+
+#. `EVENT_FRAGMENT`: Tracks various stages of directory fragmentation (split/merge).
+
+#. `EVENT_IMPORTSTART`: Logged when an MDS rank starts importing directory fragments.
+
+#. `EVENT_IMPORTFINISH`: Logged when an MDS rank finishes importing directory fragments.
+
+#. `EVENT_NOOP`: No operation event type for skipping over a journal region.
+
+#. `EVENT_OPEN`: Tracks which inodes have open file handles.
+
+#. `EVENT_RESETJOURNAL`: Used to mark a journal as `reset` post truncation.
+
+#. `EVENT_SESSION`: Tracks open client sessions.
+
+#. `EVENT_SLAVEUPDATE`: Logs various stages of an operation that has been forwarded to a (slave) mds.
+
+#. `EVENT_SUBTREEMAP`: Map of directory inodes to directory contents (subtree partition).
+
+#. `EVENT_TABLECLIENT`: Log transition states of MDSs view of client tables (snap/anchor).
+
+#. `EVENT_TABLESERVER`: Log transition states of MDSs view of server tables (snap/anchor).
+
+#. `EVENT_UPDATE`: Log file operations on an inode.
+
+#. `EVENT_SEGMENT`: Log a new journal segment boundary.
+
+#. `EVENT_LID`: Mark the beginning of a journal without a logical subtree map.
+
+Journal Segments
+----------------
+
+The MDS journal is composed of logical segments, called LogSegments in the
+code. These segments are used to collect metadata updates by multiple events
+into one logical unit for the purposes of trimming. Whenever the journal tries
+to commit metadata operations (e.g. flush a file create out as an omap update
+to a dirfrag object), it does so in a replayable batch of updates from the
+LogSegment. The updates must be replayable in case the MDS fails during the
+series of updates to various metadata objects. The reason the updates are
+performed in batch is to group updates to the same metadata object (a dirfrag)
+where multiple omap entries are probably updated in the same time period.
+
+Once a segment is trimmed, it is considered "expired". An expired segment is
+eligible for deletion by the journaler as all of its updates are flushed to the
+backing RADOS objects. This is done by updating the "expire position" of the
+journaler to advance past the end of the expired segment. Some expired segments
+may be kept in the journal to improve cache locality when the MDS restarts.
+
+For most of CephFS's history (up to 2023), the journal segments were delineated
+by subtree maps, the ``ESubtreeMap`` event. The major reason for this is that
+journal recovery must start with a copy of the subtree map before replaying any
+other events.
+
+Now, log segments can be delineated by events which are a ``SegmentBoundary``.
+These include, ``ESubtreeMap``, ``EResetJournal``, ``ESegment`` (2023), or
+``ELid`` (2023). For ``ESegment``, this light-weight segment boundary allows
+the MDS to journal the subtree map less frequently while also keeping the
+journal segments small to keep trimming events short. In order to maintain the
+constraint that the first event journal replay sees is the ``ESubtreeMap``,
+those segments beginning with that event are considered "major segments" and a
+new constraint was added to the deletion of expired segments: the first segment
+of the journal must always be a major segment.
+
+The ``ELid`` event exists to mark the MDS journal as "new" where a logical
+``LogSegment`` and log sequence number is required for other operations to
+proceed, in particular the MDSTable operations. The MDS uses this event when
+creating a rank or shutting it down. No subtree map is required when replaying
+the rank from this initial state.
+
+
+Configurations
+--------------
+
+The targetted size of a log segment in terms of number of events is controlled by:
+
+.. confval:: mds_log_events_per_segment
+
+The frequency of major segments (noted by the journaling of the latest ``ESubtreeMap``) is controlled by:
+
+.. confval:: mds_log_major_segment_event_ratio
+
+When ``mds_log_events_per_segment * mds_log_major_segment_event_ratio``
+non-``ESubtreeMap`` events are logged, the MDS will journal a new
+``ESubtreeMap``. This is necessary to allow the journal to shrink in size
+during the trimming of expired segments.
+
+The target maximum number of segments is controlled by:
+
+.. confval:: mds_log_max_segments
+
+The MDS will often sit a little above this number due to non-major segments
+awaiting trimming up to the next major segment.