mirror of
https://github.com/kemko/nomad.git
synced 2026-01-01 16:05:42 +03:00
209 lines
6.4 KiB
JavaScript
209 lines
6.4 KiB
JavaScript
/**
|
|
* Copyright (c) HashiCorp, Inc.
|
|
* SPDX-License-Identifier: MPL-2.0
|
|
*/
|
|
|
|
// @ts-check
|
|
import Component from '@glimmer/component';
|
|
import { task } from 'ember-concurrency';
|
|
import { tracked } from '@glimmer/tracking';
|
|
import { alias } from '@ember/object/computed';
|
|
import messageFromAdapterError from 'nomad-ui/utils/message-from-adapter-error';
|
|
import { jobAllocStatuses } from '../../../utils/allocation-client-statuses';
|
|
|
|
export default class JobStatusPanelDeployingComponent extends Component {
|
|
@alias('args.job') job;
|
|
@alias('args.handleError') handleError = () => {};
|
|
|
|
get allocTypes() {
|
|
return jobAllocStatuses[this.args.job.type].map((type) => {
|
|
return {
|
|
label: type,
|
|
};
|
|
});
|
|
}
|
|
|
|
@tracked oldVersionAllocBlockIDs = [];
|
|
|
|
// Called via did-insert; sets a static array of "outgoing"
|
|
// allocations we can track throughout a deployment
|
|
establishOldAllocBlockIDs() {
|
|
this.oldVersionAllocBlockIDs = this.job.allocations.filter(
|
|
(a) => a.clientStatus === 'running' && a.isOld
|
|
);
|
|
}
|
|
|
|
@task(function* () {
|
|
try {
|
|
yield this.job.latestDeployment.content.promote();
|
|
} catch (err) {
|
|
this.handleError({
|
|
title: 'Could Not Promote Deployment',
|
|
description: messageFromAdapterError(err, 'promote deployments'),
|
|
});
|
|
}
|
|
})
|
|
promote;
|
|
|
|
@task(function* () {
|
|
try {
|
|
yield this.job.latestDeployment.content.fail();
|
|
} catch (err) {
|
|
this.handleError({
|
|
title: 'Could Not Fail Deployment',
|
|
description: messageFromAdapterError(err, 'fail deployments'),
|
|
});
|
|
}
|
|
})
|
|
fail;
|
|
|
|
@alias('job.latestDeployment') deployment;
|
|
@alias('totalAllocs') desiredTotal;
|
|
|
|
get oldVersionAllocBlocks() {
|
|
return this.job.allocations
|
|
.filter((allocation) => this.oldVersionAllocBlockIDs.includes(allocation))
|
|
.reduce((alloGroups, currentAlloc) => {
|
|
const status = currentAlloc.clientStatus;
|
|
|
|
if (!alloGroups[status]) {
|
|
alloGroups[status] = {
|
|
healthy: { nonCanary: [] },
|
|
unhealthy: { nonCanary: [] },
|
|
};
|
|
}
|
|
alloGroups[status].healthy.nonCanary.push(currentAlloc);
|
|
|
|
return alloGroups;
|
|
}, {});
|
|
}
|
|
|
|
get newVersionAllocBlocks() {
|
|
let availableSlotsToFill = this.desiredTotal;
|
|
let allocationsOfDeploymentVersion = this.job.allocations.filter(
|
|
(a) => !a.isOld
|
|
);
|
|
|
|
let allocationCategories = this.allocTypes.reduce((categories, type) => {
|
|
categories[type.label] = {
|
|
healthy: { canary: [], nonCanary: [] },
|
|
unhealthy: { canary: [], nonCanary: [] },
|
|
};
|
|
return categories;
|
|
}, {});
|
|
|
|
for (let alloc of allocationsOfDeploymentVersion) {
|
|
if (availableSlotsToFill <= 0) {
|
|
break;
|
|
}
|
|
let status = alloc.clientStatus;
|
|
let canary = alloc.isCanary ? 'canary' : 'nonCanary';
|
|
|
|
// Health status only matters in the context of a "running" allocation.
|
|
// However, healthy/unhealthy is never purged when an allocation moves to a different clientStatus
|
|
// Thus, we should only show something as "healthy" in the event that it is running.
|
|
// Otherwise, we'd have arbitrary groupings based on previous health status.
|
|
let health =
|
|
status === 'running'
|
|
? alloc.isHealthy
|
|
? 'healthy'
|
|
: 'unhealthy'
|
|
: 'unhealthy';
|
|
|
|
if (allocationCategories[status]) {
|
|
// If status is failed or lost, we only want to show it IF it's used up its restarts/rescheds.
|
|
// Otherwise, we'd be showing an alloc that had been replaced.
|
|
if (alloc.willNotRestart) {
|
|
if (!alloc.willNotReschedule) {
|
|
// Dont count it
|
|
continue;
|
|
}
|
|
}
|
|
allocationCategories[status][health][canary].push(alloc);
|
|
availableSlotsToFill--;
|
|
}
|
|
}
|
|
|
|
// Fill unplaced slots if availableSlotsToFill > 0
|
|
if (availableSlotsToFill > 0) {
|
|
allocationCategories['unplaced'] = {
|
|
healthy: { canary: [], nonCanary: [] },
|
|
unhealthy: { canary: [], nonCanary: [] },
|
|
};
|
|
allocationCategories['unplaced']['healthy']['nonCanary'] = Array(
|
|
availableSlotsToFill
|
|
)
|
|
.fill()
|
|
.map(() => {
|
|
return { clientStatus: 'unplaced' };
|
|
});
|
|
}
|
|
|
|
return allocationCategories;
|
|
}
|
|
|
|
get newRunningHealthyAllocBlocks() {
|
|
return [
|
|
...this.newVersionAllocBlocks['running']['healthy']['canary'],
|
|
...this.newVersionAllocBlocks['running']['healthy']['nonCanary'],
|
|
];
|
|
}
|
|
|
|
get rescheduledAllocs() {
|
|
return this.job.allocations.filter((a) => !a.isOld && a.hasBeenRescheduled);
|
|
}
|
|
|
|
get restartedAllocs() {
|
|
return this.job.allocations.filter((a) => !a.isOld && a.hasBeenRestarted);
|
|
}
|
|
|
|
// #region legend
|
|
get newAllocsByStatus() {
|
|
return Object.entries(this.newVersionAllocBlocks).reduce(
|
|
(counts, [status, healthStatusObj]) => {
|
|
counts[status] = Object.values(healthStatusObj)
|
|
.flatMap((canaryStatusObj) => Object.values(canaryStatusObj))
|
|
.flatMap((canaryStatusArray) => canaryStatusArray).length;
|
|
return counts;
|
|
},
|
|
{}
|
|
);
|
|
}
|
|
|
|
get newAllocsByCanary() {
|
|
return Object.values(this.newVersionAllocBlocks)
|
|
.flatMap((healthStatusObj) => Object.values(healthStatusObj))
|
|
.flatMap((canaryStatusObj) => Object.entries(canaryStatusObj))
|
|
.reduce((counts, [canaryStatus, items]) => {
|
|
counts[canaryStatus] = (counts[canaryStatus] || 0) + items.length;
|
|
return counts;
|
|
}, {});
|
|
}
|
|
|
|
get newAllocsByHealth() {
|
|
return {
|
|
healthy: this.newRunningHealthyAllocBlocks.length,
|
|
'health unknown':
|
|
this.totalAllocs - this.newRunningHealthyAllocBlocks.length,
|
|
};
|
|
}
|
|
// #endregion legend
|
|
|
|
get oldRunningHealthyAllocBlocks() {
|
|
return this.oldVersionAllocBlocks.running?.healthy?.nonCanary || [];
|
|
}
|
|
get oldCompleteHealthyAllocBlocks() {
|
|
return this.oldVersionAllocBlocks.complete?.healthy?.nonCanary || [];
|
|
}
|
|
|
|
// TODO: eventually we will want this from a new property on a job.
|
|
// TODO: consolidate w/ the one in steady.js
|
|
get totalAllocs() {
|
|
// v----- Experimental method: Count all allocs. Good for testing but not a realistic representation of "Desired"
|
|
// return this.allocTypes.reduce((sum, type) => sum + this.args.job[type.property], 0);
|
|
|
|
// v----- Realistic method: Tally a job's task groups' "count" property
|
|
return this.args.job.taskGroups.reduce((sum, tg) => sum + tg.count, 0);
|
|
}
|
|
}
|