Files
nomad/command/monitor_test.go
2024-12-19 15:41:11 +00:00

385 lines
9.4 KiB
Go

// Copyright (c) HashiCorp, Inc.
// SPDX-License-Identifier: BUSL-1.1
package command
import (
"fmt"
"strings"
"testing"
"time"
"github.com/hashicorp/cli"
"github.com/hashicorp/nomad/api"
"github.com/hashicorp/nomad/ci"
"github.com/hashicorp/nomad/helper/pointer"
"github.com/hashicorp/nomad/nomad/structs"
"github.com/shoenig/test/must"
"github.com/shoenig/test/wait"
)
func TestMonitor_Update_Eval(t *testing.T) {
ci.Parallel(t)
ui := cli.NewMockUi()
mon := newMonitor(ui, nil, fullId)
// Evals triggered by jobs log
state := &evalState{
status: structs.EvalStatusPending,
job: "job1",
}
mon.update(state)
out := ui.OutputWriter.String()
if !strings.Contains(out, "job1") {
t.Fatalf("missing job\n\n%s", out)
}
ui.OutputWriter.Reset()
// Evals triggered by nodes log
state = &evalState{
status: structs.EvalStatusPending,
node: "12345678-abcd-efab-cdef-123456789abc",
}
mon.update(state)
out = ui.OutputWriter.String()
if !strings.Contains(out, "12345678-abcd-efab-cdef-123456789abc") {
t.Fatalf("missing node\n\n%s", out)
}
// Transition to pending should not be logged
if strings.Contains(out, structs.EvalStatusPending) {
t.Fatalf("should skip status\n\n%s", out)
}
ui.OutputWriter.Reset()
// No logs sent if no update
mon.update(state)
if out := ui.OutputWriter.String(); out != "" {
t.Fatalf("expected no output\n\n%s", out)
}
// Status change sends more logs
state = &evalState{
status: structs.EvalStatusComplete,
node: "12345678-abcd-efab-cdef-123456789abc",
}
mon.update(state)
out = ui.OutputWriter.String()
if !strings.Contains(out, structs.EvalStatusComplete) {
t.Fatalf("missing status\n\n%s", out)
}
}
func TestMonitor_Update_Allocs(t *testing.T) {
ci.Parallel(t)
ui := cli.NewMockUi()
mon := newMonitor(ui, nil, fullId)
// New allocations write new logs
state := &evalState{
allocs: map[string]*allocState{
"alloc1": {
id: "87654321-abcd-efab-cdef-123456789abc",
group: "group1",
node: "12345678-abcd-efab-cdef-123456789abc",
desired: structs.AllocDesiredStatusRun,
client: structs.AllocClientStatusPending,
index: 1,
},
},
}
mon.update(state)
// Logs were output
out := ui.OutputWriter.String()
if !strings.Contains(out, "87654321-abcd-efab-cdef-123456789abc") {
t.Fatalf("missing alloc\n\n%s", out)
}
if !strings.Contains(out, "group1") {
t.Fatalf("missing group\n\n%s", out)
}
if !strings.Contains(out, "12345678-abcd-efab-cdef-123456789abc") {
t.Fatalf("missing node\n\n%s", out)
}
if !strings.Contains(out, "created") {
t.Fatalf("missing created\n\n%s", out)
}
ui.OutputWriter.Reset()
// No change yields no logs
mon.update(state)
if out := ui.OutputWriter.String(); out != "" {
t.Fatalf("expected no output\n\n%s", out)
}
ui.OutputWriter.Reset()
// Alloc updates cause more log lines
state = &evalState{
allocs: map[string]*allocState{
"alloc1": {
id: "87654321-abcd-efab-cdef-123456789abc",
group: "group1",
node: "12345678-abcd-efab-cdef-123456789abc",
desired: structs.AllocDesiredStatusRun,
client: structs.AllocClientStatusRunning,
index: 2,
},
},
}
mon.update(state)
// Updates were logged
out = ui.OutputWriter.String()
if !strings.Contains(out, "87654321-abcd-efab-cdef-123456789abc") {
t.Fatalf("missing alloc\n\n%s", out)
}
if !strings.Contains(out, "pending") {
t.Fatalf("missing old status\n\n%s", out)
}
if !strings.Contains(out, "running") {
t.Fatalf("missing new status\n\n%s", out)
}
}
func TestMonitor_Update_AllocModification(t *testing.T) {
ci.Parallel(t)
ui := cli.NewMockUi()
mon := newMonitor(ui, nil, fullId)
// New allocs with a create index lower than the
// eval create index are logged as modifications
state := &evalState{
index: 2,
allocs: map[string]*allocState{
"alloc3": {
id: "87654321-abcd-bafe-cdef-123456789abc",
node: "12345678-abcd-efab-cdef-123456789abc",
group: "group2",
index: 1,
},
},
}
mon.update(state)
// Modification was logged
out := ui.OutputWriter.String()
if !strings.Contains(out, "87654321-abcd-bafe-cdef-123456789abc") {
t.Fatalf("missing alloc\n\n%s", out)
}
if !strings.Contains(out, "group2") {
t.Fatalf("missing group\n\n%s", out)
}
if !strings.Contains(out, "12345678-abcd-efab-cdef-123456789abc") {
t.Fatalf("missing node\n\n%s", out)
}
if !strings.Contains(out, "modified") {
t.Fatalf("missing modification\n\n%s", out)
}
}
func TestMonitor_Monitor(t *testing.T) {
ci.Parallel(t)
srv, client, _ := testServer(t, false, nil)
defer srv.Shutdown()
// Create the monitor
ui := cli.NewMockUi()
mon := newMonitor(ui, client, fullId)
// Submit a job - this creates a new evaluation we can monitor
job := testJob("job1")
resp, _, err := client.Jobs().Register(job, nil)
if err != nil {
t.Fatalf("err: %s", err)
}
// Start monitoring the eval
var code int
doneCh := make(chan struct{})
go func() {
defer close(doneCh)
code = mon.monitor(resp.EvalID)
}()
// Wait for completion
select {
case <-doneCh:
case <-time.After(5 * time.Second):
t.Fatalf("eval monitor took too long")
}
// Check the return code. We should get exit code 2 as there
// would be a scheduling problem on the test server (no clients).
if code != 2 {
t.Fatalf("expect exit 2, got: %d", code)
}
// Check the output
out := ui.OutputWriter.String()
if !strings.Contains(out, resp.EvalID) {
t.Fatalf("missing eval\n\n%s", out)
}
if !strings.Contains(out, "finished with status") {
t.Fatalf("missing final status\n\n%s", out)
}
}
func TestMonitor_MonitorBlockedEval(t *testing.T) {
ci.Parallel(t)
srv, client, _ := testServer(t, false, nil)
defer srv.Shutdown()
ui := cli.NewMockUi()
mon := newMonitor(ui, client, fullId)
// Submit a service job.
// Since there are no clients this will create a blocked eval.
job := testJob("job1")
job.Type = pointer.Of("service")
job.TaskGroups[0].Tasks[0].Config["run_for"] = "300s"
resp, _, err := client.Jobs().Register(job, nil)
must.NoError(t, err)
// Start monitoring the eval and collect return code.
var code int
doneCh := make(chan struct{})
go func() {
defer close(doneCh)
code = mon.monitor(resp.EvalID)
}()
// Helper function to check for eval and deployment status.
waitForEvalDeploymentStatusFunc := func(evalStatus string, deploymentStatus string) func() error {
return func() error {
// Verify most recent job eval is blocked.
evals, _, err := client.Jobs().Evaluations(*job.ID, nil)
if err != nil {
return fmt.Errorf("failed to fetch job evals: %w", err)
}
if len(evals) < 1 {
return fmt.Errorf("expected at least one eval, got %d", len(evals))
}
eval := evals[0]
if eval.Status != evalStatus {
return fmt.Errorf("expected eval to be %q, got %q", evalStatus, eval.Status)
}
// Verify job deployment is running.
ds, _, err := client.Jobs().Deployments(*job.ID, false, nil)
if err != nil {
return fmt.Errorf("failed to fetch job deployments: %w", err)
}
if len(ds) != 1 {
return fmt.Errorf("expected 1 deployment, found %d", len(ds))
}
d := ds[0]
if d.Status != deploymentStatus {
return fmt.Errorf("expected deployment to be %q, got %q", deploymentStatus, d.Status)
}
return nil
}
}
// Wait until job eval is blocked and deployment is running.
must.Wait(t,
wait.InitialSuccess(
wait.ErrorFunc(waitForEvalDeploymentStatusFunc(
api.EvalStatusBlocked,
api.DeploymentStatusRunning,
)),
wait.Timeout(3*time.Second),
wait.Gap(1*time.Second),
),
must.Sprintf("failed to wait for blocked deployment"),
)
// Add client to provide the necessary capacity and unblock the eval.
srvRPCAddr := srv.GetConfig().AdvertiseAddrs.RPC
testClient(t, "client1", newClientAgentConfigFunc("global", "classA", srvRPCAddr))
// Wait until job eval is complete and deployment is successful.
must.Wait(t,
wait.InitialSuccess(
wait.ErrorFunc(waitForEvalDeploymentStatusFunc(
api.EvalStatusComplete,
api.DeploymentStatusSuccessful,
)),
wait.Timeout(30*time.Second),
wait.Gap(1*time.Second),
),
must.Sprintf("failed to wait for blocked deployment"),
)
// Wait for command to complete.
select {
case <-doneCh:
case <-time.After(5 * time.Second):
t.Fatalf("eval monitor took too long")
}
// Verify status code is 0 since deployment was succesful.
must.Zero(t, code)
}
func TestMonitor_formatAllocMetric(t *testing.T) {
ci.Parallel(t)
tests := []struct {
Name string
Metrics *api.AllocationMetric
Expected string
}{
{
Name: "display all possible scores",
Metrics: &api.AllocationMetric{
NodesEvaluated: 3,
NodesInPool: 3,
ScoreMetaData: []*api.NodeScoreMeta{
{
NodeID: "node-1",
Scores: map[string]float64{
"score-1": 1,
"score-2": 2,
},
NormScore: 1,
},
{
NodeID: "node-2",
Scores: map[string]float64{
"score-1": 1,
"score-3": 3,
},
NormScore: 2,
},
{
NodeID: "node-3",
Scores: map[string]float64{
"score-4": 4,
},
NormScore: 3,
},
},
},
Expected: `
Node score-1 score-2 score-3 score-4 final score
node-1 1 2 0 0 1
node-2 1 0 3 0 2
node-3 0 0 0 4 3
`,
},
}
for _, tc := range tests {
t.Run(tc.Name, func(t *testing.T) {
got := formatAllocMetrics(tc.Metrics, true, "")
must.Eq(t, strings.TrimSpace(tc.Expected), got)
})
}
}