From 4075b0b8baaf3dfac01a69221653471abb716540 Mon Sep 17 00:00:00 2001 From: Aimee Ukasick Date: Mon, 28 Apr 2025 08:37:23 -0500 Subject: [PATCH] Docs: Add garbage collection page (#25715) * add garbage collection page * finish client; add resources section * finish server section; task driver section * add front matter description * fix typos * Address Tim's feedback --- .../docs/operations/garbage-collection.mdx | 214 ++++++++++++++++++ website/data/docs-nav-data.json | 4 + 2 files changed, 218 insertions(+) create mode 100644 website/content/docs/operations/garbage-collection.mdx diff --git a/website/content/docs/operations/garbage-collection.mdx b/website/content/docs/operations/garbage-collection.mdx new file mode 100644 index 000000000..3fc2a5b84 --- /dev/null +++ b/website/content/docs/operations/garbage-collection.mdx @@ -0,0 +1,214 @@ +--- +layout: docs +page_title: Garbage collection +description: |- + Nomad garbage collects Access Control List (ACL) tokens, allocations, deployments, encryption root keys, evaluations, jobs, nodes, plugins, and Container Storage Interface (CSI) volumes. Learn about server-side and client-side garbage collection processes, including configuration and triggers. +--- + +# Garbage collection + +Nomad garbage collection is not the same as garbage collection in a programming +language, but the motivation behind its design is similar: garbage collection +frees up memory allocated for objects that the schedular no longer references or +needs. Nomad only garbage collects objects that are in a terminal state and only +after a delay to allow inspection or debugging. + +Nomad runs garbage collection processes on servers and on client nodes. You may +also manually trigger garbage collection on the server. + +Nomad garbage collects the following objects: + +- [ACL token](#configuration) +- [Allocation](#client-side-garbage-collection) +- [CSI Plugin](#configuration) +- [Deployment](#configuration) +- [Encryption root key](#configuration) +- [Evaluation](#configuration) +- [Job](#configuration) +- [Node](#configuration) +- [Volume](#configuration) + +## Cascading garbage collection + +Nomad's scheduled garbage collection processes generally handle each resource +type independently. However, there is an implicit cascading relationship because +of how objects reference each other. In practice, when Nomad garbage collects a +higher-level object, Nomad also removes the object's associated sub-objects to +prevent orphaned objects. + +For example, garbage collecting a job also causes Nomad to drop all of that +job's remaining evaluations, deployments, and allocation records from the state. +Nomad garbage collects those objects, either as part of the job garbage +collection process or by each object's own garbage collection processes running +immediately after. Nomad's scheduled garbage collection processes only garbage +collect objects after they are terminal for at least the specified time +threshold and no longer needed for future scheduling decisions. Note that when +you force garbage collection by running the `nomad system gc` command, Nomad +ignores the specified time threshold. + +## Server-side garbage collection + +The Nomad server leader starts periodic garbage collection processes that clean +objects marked for garbage collection from memory. Nomad automatically marks +some objects, like evaluations, for garbage collection. Alternatively, you may +manually mark jobs for garbage collection by running `nomad system gc`, which +runs the garbage collection process. + +### Configuration + +These settings govern garbage collection behavior on the server nodes. You may +review the intervals in the [`config.go` +class](https://github.com/hashicorp/nomad/blob/b11619010e1c83488e14e2785569e515b2769062/nomad/config.go#L564) +for objects without a configurable interval setting. + +| Object | Interval | Threshold | +|---|---|---| +| **ACL token** | 5 minutes | [`acl_token_gc_threshold`](/nomad/docs/configuration/server#acl_token_gc_threshold)
Default: 1 hour | +| **CSI Plugin** | 5 minutes | [`csi_plugin_gc_threshold`](/nomad/docs/configuration/server#csi_plugin_gc_threshold)
Default: 1 hour | +| **Deployment** | 5 minutes | [`deployment_gc_threshold`](/nomad/docs/configuration/server#deployment_gc_threshold)
Default: 1 hour | +| **Encryption root key** | [`root_key_gc_interval`](/nomad/docs/configuration/server#root_key_gc_interval)
Default: 10 minutes | [`root_key_gc_threshold`](/nomad/docs/configuration/server#root_key_gc_threshold)
Default: 1 hour | +| **Evaluation** | 5 minutes | [`eval_gc_threshold`](/nomad/docs/configuration/server#eval_gc_threshold)
Default: 1 hour | +| **Evaluation, batch** | 5 minutes | [`batch_eval_gc_threshold`](/nomad/docs/configuration/server#batch_eval_gc_threshold)
Default: 24 hours | +| **Job** | [`job_gc_interval`](/nomad/docs/configuration/server#job_gc_interval)
Default: 5 minutes | [`job_gc_threshold`](/nomad/docs/configuration/server#job_gc_threshold)
Default: 4 hours | +| **Node** | 5 minutes | [`node_gc_threshold`](/nomad/docs/configuration/server#node_gc_threshold)
Default: 24 hours | +| **Volume** | [`csi_volume_claim_gc_interval`](/nomad/docs/configuration/server#csi_volume_claim_gc_interval)
Default: 5 minutes| [`csi_volume_claim_gc_threshold`](/nomad/docs/configuration/server#csi_volume_claim_gc_threshold)
Default: 1 hour | + +### Triggers + +The server garbage collection processes wake up at configured intervals to scan +for any expired or terminal objects to permanently delete, provided the object's +time in a terminal state exceeds its garbage collection threshold. For example, +a job's default garbage collection threshold is four hours, so the job must be +in a terminal state for at least four hours before the garbage collection +process permanently deletes the job and its dependent objects. + +When you force garbage collection by manually running the `nomad system gc` +command, you are telling the garbage collection process to ignore thresholds and +immediately purge all terminal objects on all servers and clients. + +## Client-side garbage collection + +On each client node, Nomad must clean up resources from terminated allocations +to free disk and memory on the machine. + +### Configuration + +These settings govern allocation garbage collection behavior on each client node. + +| Parameter | Default | Description | +| -------- | ------- | ------------- | +| [`gc_interval`](/nomad/docs/configuration/client#gc_interval) | 1 minute | Interval at which Nomad attempts to garbage collect terminal allocation directories | +| [ `gc_disk_usage_threshold` ](/nomad/docs/configuration/client#gc_disk_usage_threshold) | 80 | Disk usage percent which Nomad tries to maintain by garbage collecting terminal allocations | +| [ `gc_inode_usage_threshold` ](/nomad/docs/configuration/client#gc_inode_usage_threshold) | 70 | Inode usage percent which Nomad tries to maintain by garbage collecting terminal allocations | +| [ `gc_max_allocs` ](/nomad/docs/configuration/client#gc_max_allocs) | 50 | Maximum number of allocations which a client will track before triggering a garbage collection of terminal allocations | +| [ `gc_parallel_destroys ` ](/nomad/docs/configuration/client#gc_parallel_destroys) | 2 | Maximum number of parallel destroys allowed by the garbage collector | + +Refer to the [client block in agent configuration +reference](/nomad/docs/configuration/client) for complete parameter descriptions +and examples. + +Note that there is no time-based retention setting for allocations. Unlike jobs +or evaluations, you cannot specify a time to keep allocations alive before +garbage collection. As soon as an allocation is terminal, it becomes eligible +for cleanup if the configured thresholds demand it. + +### Triggers + +Nomad's client runs allocation garbage collection based on these triggers: + +- Scheduled interval + + The garbage collection process launches a ticker based on the configured + `gc_interval`. On each tick, the garbage collection process checks to see if it needs to remove terminal allocations. + +- Terminal state + + When an allocation transitions to a terminal state, Nomad marks + the allocation for garbage collection and then signals the garbage collection + process to run immediately. + +- Allocation placement + + Nomad may preemptively run garbage collection to make room for new + allocations. The client garbage collects older, terminal allocations if adding new allocations would exceed the `gc_max_allocs` limit. + +- Forced garbage collection + + When you force garbage collection by running the `nomad system gc` command, + the garbage collection process removes all terminal objects on all servers and + clients, ignoring thresholds. + +Nomad does not continuously monitor disk or inode usage to trigger garbage +collection. Instead, Nomad only checks disk and inode thresholds when one of the +aforementioned triggers invokes the garbage collection process. The +`gc_inode_usage_threshold` and `gc_disk_usage_threshold` values do not trigger +garbage collection; rather, those values influence how the garbage collector +behaves during a collection run. + +### Allocation selection + +When the garbage collection process runs, Nomad destroys as many finished +allocations as needed to meet the resource thresholds. The client maintains a +priority queue of terminal allocations ordered by the time they were marked +finished, oldest first. + +The process repeatedly evicts allocations from the queue until the conditions +are back within configured limits. Specifically, the garbage collection loop +checks, in order: + +1. If disk usage exceeds `gc_disk_usage_threshold` value +1. If inode usage exceeds `gc_inode_usage_threshold` value +1. If the count of allocations exceeds `gc_max_allocs` value + +If any one of these conditions is true, the garbage collector selects the oldest +finished allocation for removal. + +After deleting one allocation, the loop re-checks the metrics and continues +removing the next-oldest allocation until all thresholds are satisfied or +until there are no more terminal allocs. This means in a single run, the +garbage collection removes multiple allocations back-to-back if the node was +far over the limits. The evictions happen in termination-time order, which is +oldest completed allocations first. + +If node's usage and allocation count are under the limits, a normal garbage +collection cycle does not remove any allocations. In other words, periodic and +event-driven garbage collection does not delete allocations just because they +are finished. There has to be pressure or a limit reached. The exception is when +an administrative command or server-side removal triggers client-side garbage +collection. Aside from that forced scenario, the default behavior is +threshold-driven: Nomad leaves allocations on disk until it needs to reclaim +those allocations due to space, inode, or count limits being hit. + +### Task driver resources garbage collection + +Most task drivers do not have their own garbage collection process. When an +allocation is terminal, the client garbage collection process communicates with +the task driver to ensure the task's resources have been cleaned up. Note that +the Docker task driver periodically cleans up its own resources. Refer to the +[Docker task driver plugin +options](https://developer.hashicorp.com/nomad/docs/drivers/docker#gc) for +details. + +When a task has configured restart attempts and the task fails, the Nomad client +attempts an in-place task restart within the same allocation. The task driver +starts a new process or container for the task. If the task continues to fail +and exceeds the configured restart attempts, Nomad terminates the task and marks +the allocation as terminal. The task driver then cleans up its resources, such +as a Docker container or cgroups. When the garbage collection process runs, it +makes sure that the task driver cleanup is done before deleting the allocation. +If a task driver fails to clean up properly, Nomad logs errors but continues the +garbage collection process. Task driver cleanup failure issues can influence +when the allocation truly frees up. For instance, if volumes are not detached, +disk space might not be fully reclaimed until fixed. + +## Resources + +- [Nomad's internal garbage collection and optimization discovery during the + Nomad Bench project blog post](https://www.hashicorp.com/en/blog/nomad-garbage-collection-optimization-discovery-during-nomad-bench) +- Configuration + + - [client Block in Agent Configuration](/nomad/docs/configuration/client) + - [server Block in Agent Configuration](/nomad/docs/configuration/server) + +- [the `nomad system gc` command reference](/nomad/docs/commands/system/gc) +- [System HTTP API Force GC](/nomad/api-docs/system#force-gc) diff --git a/website/data/docs-nav-data.json b/website/data/docs-nav-data.json index 3312a6792..c4e6a7a20 100644 --- a/website/data/docs-nav-data.json +++ b/website/data/docs-nav-data.json @@ -2393,6 +2393,10 @@ { "title": "IPv6 Support", "path": "operations/ipv6-support" + }, + { + "title": "Garbage Collection", + "path": "operations/garbage-collection" } ] },