Files
nomad/website/content/docs/job-scheduling/affinity.mdx
Aimee Ukasick 53b083b8c5 Docs: Nomad IA (#26063)
* Move commands from docs to its own root-level directory

* temporarily use modified dev-portal branch with nomad ia changes

* explicitly clone nomad ia exp branch

* retrigger build, fixed dev-portal broken build

* architecture, concepts and get started individual pages

* fix get started section destinations

* reference section

* update repo comment in website-build.sh to show branch

* docs nav file update capitalization

* update capitalization to force deploy

* remove nomad-vs-kubernetes dir; move content to what is nomad pg

* job section

* Nomad operations category, deploy section

* operations category, govern section

* operations - manage

* operations/scale; concepts scheduling fix

* networking

* monitor

* secure section

* remote auth-methods folder and move up pages to sso; linkcheck

* Fix install2deploy redirects

* fix architecture redirects

* Job section: Add missing section index pages

* Add section index pages so breadcrumbs build correctly

* concepts/index fix front matter indentation

* move task driver plugin config to new deploy section

* Finish adding full URL to tutorials links in nav

* change SSO to Authentication in nav and file system

* Docs NomadIA: Move tutorials into NomadIA branch (#26132)

* Move governance and policy from tutorials to docs

* Move tutorials content to job-declare section

* run jobs section

* stateful workloads

* advanced job scheduling

* deploy section

* manage section

* monitor section

* secure/acl and secure/authorization

* fix example that contains an unseal key in real format

* remove images from sso-vault

* secure/traffic

* secure/workload-identities

* vault-acl change unseal key and root token in command output sample

* remove lines from sample output

* fix front matter

* move nomad pack tutorials to tools

* search/replace /nomad/tutorials links

* update acl overview with content from deleted architecture/acl

* fix spelling mistake

* linkcheck - fix broken links

* fix link to Nomad variables tutorial

* fix link to Prometheus tutorial

* move who uses Nomad to use cases page; move spec/config shortcuts

add dividers

* Move Consul out of Integrations; move namespaces to govern

* move integrations/vault to secure/vault; delete integrations

* move ref arch to docs; rename Deploy Nomad back to Install Nomad

* address feedback

* linkcheck fixes

* Fixed raw_exec redirect

* add info from /nomad/tutorials/manage-jobs/jobs

* update page content with newer tutorial

* link updates for architecture sub-folders

* Add redirects for removed section index pages. Fix links.

* fix broken links from linkcheck

* Revert to use dev-portal main branch instead of nomadIA branch

* build workaround: add intro-nav-data.json with single entry

* fix content-check error

* add intro directory to get around Vercel build error

* workound for emtpry directory

* remove mdx from /intro/ to fix content-check and git snafu

* Add intro index.mdx so Vercel build should work

---------

Co-authored-by: Tu Nguyen <im2nguyen@gmail.com>
2025-07-08 19:24:52 -05:00

251 lines
9.0 KiB
Plaintext

---
layout: docs
page_title: Job placements with affinities
description: >-
Configure affinities to express placement preferences for your jobs. Create a
job with an affinity, submit it to Nomad, and monitor it after placement.
---
# Job placements with affinities
The [affinity][affinity-stanza] stanza allows operators to express placement
preferences for their jobs on particular types of nodes. Note that there is a
key difference between the [constraint][constraint] stanza and the affinity
stanza. The constraint stanza strictly filters where jobs are run based on
[attributes][attributes] and [client metadata][client-metadata]. If no nodes are
found to match, the placement does not succeed. The affinity stanza acts like a
"soft constraint." Nomad will attempt to match the desired affinity, but
placement will succeed even if no nodes match the desired criteria. This is done
in conjunction with scoring based on the Nomad scheduler's bin packing algorithm
which you can read more about [here][scheduling].
In this guide, you will encounter a sample application. Your application can run
in datacenters `dc1` and `dc2`; however, you have a strong preference to run it in
dc2. You will learn how to configure the job to inform the scheduler of your
preference, while still allowing it to place your workload in `dc1` if the
desired resources aren't available in dc2.
By specify an affinity with the proper [weight], the Nomad scheduler can find
the best nodes on which to place your job. The affinity weight will be included
when scoring nodes for placement along with other factors like the bin-packing
algorithm.
### Prerequisites
To perform the tasks described in this guide, you need to have a Nomad
environment with Consul installed. You can use this [repository] to provision a
sandbox environment. This guide will assume a cluster with one server node and
three client nodes.
<Tip>
This guide is for demo purposes and is only using a single
server node. In a production cluster, 3 or 5 server nodes are recommended.
</Tip>
## Place one of the client nodes in a different datacenter
You are going express your job placement preference based on the datacenter your
nodes are located in. Choose one of your client nodes and edit
`/etc/nomad.d/nomad.hcl` to change its location to `dc2`. A snippet of an
example configuration file is show below with the required change is shown
below.
```hcl
data_dir = "/opt/nomad/data"
bind_addr = "0.0.0.0"
datacenter = "dc2"
# Enable the client
client {
enabled = true
# ...
}
```
After making the change on your chosen client node, restart the Nomad service
```shell-session
$ sudo systemctl restart nomad
```
If everything worked correctly, one of your nodes will now show datacenter `dc2`
when you run the [`nomad node status`][node-status] command.
```shell-session
$ nomad node status
ID DC Name Class Drain Eligibility Status
3592943e dc1 ip-172-31-27-159 <none> false eligible ready
3dea0188 dc1 ip-172-31-16-175 <none> false eligible ready
6b6e9518 dc2 ip-172-31-27-25 <none> false eligible ready
```
## Create a job with an affinity
Create a file with the name `redis.nomad.hcl` and place the following content in it:
```hcl
job "redis" {
datacenters = ["dc1", "dc2"]
type = "service"
affinity {
attribute = "${node.datacenter}"
value = "dc2"
weight = 100
}
group "cache1" {
count = 4
network {
port "db" {
to = 6379
}
}
service {
name = "redis-cache"
port = "db"
check {
name = "alive"
type = "tcp"
interval = "10s"
timeout = "2s"
}
}
task "redis" {
driver = "docker"
config {
image = "redis:latest"
ports = ["db"]
}
}
}
}
```
Observe that the job uses the `affinity` stanza and specifies `dc2` as the value
for the `${node.datacenter}` [attribute][attributes]. It also uses the value
`100` for the [weight][weight] which will cause the Nomad scheduler to rank
nodes in datacenter `dc2` with a higher score. Keep in mind that weights can
range from -100 to 100, inclusive. Negative weights serve as anti-affinities
which cause Nomad to avoid placing allocations on nodes that match the criteria.
## Register the redis Nomad job
Run the Nomad job with the following command:
```shell-session
$ nomad run redis.nomad.hcl
==> Monitoring evaluation "11388ef2"
Evaluation triggered by job "redis"
Allocation "0dfcf0ba" created: node "6b6e9518", group "cache1"
Allocation "89a9aae9" created: node "3592943e", group "cache1"
Allocation "9a00f742" created: node "6b6e9518", group "cache1"
Allocation "fc0f21bc" created: node "3dea0188", group "cache1"
Evaluation status changed: "pending" -> "complete"
==> Evaluation "11388ef2" finished with status "complete"
```
Note that two of the allocations in this example have been placed on node
`6b6e9518`. This is the node configured to be in datacenter `dc2`. The Nomad
scheduler selected this node because of the affinity specified. All of the
allocations have not been placed on this node because the Nomad scheduler
considers other factors in the scoring such as bin-packing. This helps avoid
placing too many instances of the same job on a node and prevents reduced
capacity during a node level failure. You will take a detailed look at the
scoring in the next few steps.
## Check the status of the job
At this point, Check the status of the job and verify where the allocations
have been placed. Run the following command:
```shell-session
$ nomad status redis
```
There should be four instances of the job running in the `Summary` section of
the output as shown below:
```plaintext
...
Summary
Task Group Queued Starting Running Failed Complete Lost
cache1 0 0 4 0 0 0
Allocations
ID Node ID Task Group Version Desired Status Created Modified
0dfcf0ba 6b6e9518 cache1 0 run running 1h44m ago 1h44m ago
89a9aae9 3592943e cache1 0 run running 1h44m ago 1h44m ago
9a00f742 6b6e9518 cache1 0 run running 1h44m ago 1h44m ago
fc0f21bc 3dea0188 cache1 0 run running 1h44m ago 1h44m ago
```
You can cross-check this output with the results of the `nomad node status`
command to verify that the majority of your workload has been placed on the node
in `dc2`. In the case of the above output, that node is `6b6e9518`.
## Obtain detailed scoring information on job placement
The Nomad scheduler will not always place all of your workload on nodes you have
specified in the `affinity` stanza even if the resources are available. This is
because affinity scoring is combined with other metrics as well before making a
scheduling decision. In this step, you will take a look at some of those other
factors.
Using the output from the previous step, find an allocation that has been placed
on a node in `dc2` and use the [`nomad alloc status`][alloc status] command with
the [`-verbose`][verbose] option to obtain detailed scoring information on it.
In this example, the allocation ID to be inspected is `0dfcf0ba` (your
allocation IDs will be different).
```shell-session
$ nomad alloc status -verbose 0dfcf0ba
```
The resulting output will show the `Placement Metrics` section at the bottom.
```plaintext
...
Placement Metrics
Node binpack job-anti-affinity node-reschedule-penalty node-affinity final score
6b6e9518-d2a4-82c8-af3b-6805c8cdc29c 0.33 0 0 1 0.665
3dea0188-ae06-ad98-64dd-a761ab2b1bf3 0.33 0 0 0 0.33
3592943e-67e4-461f-d888-d5842372a4d4 0.33 0 0 0 0.33
```
Note that the results from the `binpack`, `job-anti-affinity`,
`node-reschedule-penalty`, and `node-affinity` columns are combined to produce
the numbers listed in the `final score` column for each node. The Nomad
scheduler uses the final score for each node in deciding where to make
placements.
## Next steps
Experiment with the weight provided in the `affinity` stanza (the value can be
from -100 through 100) and observe how the final score given to each node
changes (use the `nomad alloc status` command as shown in the previous step).
### Reference material
- The [affinity][affinity-stanza] stanza documentation
- [Scheduling][scheduling] with Nomad
[affinity-stanza]: /nomad/docs/job-specification/affinity
[alloc status]: /nomad/commands/alloc/status
[attributes]: /nomad/docs/reference/runtime-variable-interpolation#node-attributes
[constraint]: /nomad/docs/job-specification/constraint
[client-metadata]: /nomad/docs/configuration/client#meta
[node-status]: /nomad/commands/node/status
[scheduling]: /nomad/docs/concepts/scheduling/how-scheduling-works
[verbose]: /nomad/commands/alloc/status#verbose
[weight]: /nomad/docs/job-specification/affinity#weight
[repository]: https://github.com/hashicorp/nomad/tree/master/terraform#provision-a-nomad-cluster-in-the-cloud