4fbcc668d0
This commit introduces support for configuring mount propagation when mounting volumes with the `volume_mount` stanza on Linux targets. Similar to Kubernetes, we expose 3 options for configuring mount propagation: - private, which is equivalent to `rprivate` on Linux, which does not allow the container to see any new nested mounts after the chroot was created. - host-to-task, which is equivalent to `rslave` on Linux, which allows new mounts that have been created _outside of the container_ to be visible inside the container after the chroot is created. - bidirectional, which is equivalent to `rshared` on Linux, which allows both the container to see new mounts created on the host, but importantly _allows the container to create mounts that are visible in other containers an don the host_ private and host-to-task are safe, but bidirectional mounts can be dangerous, as if the code inside a container creates a mount, and does not clean it up before tearing down the container, it can cause bad things to happen inside the kernel. To add a layer of safety here, we require that the user has ReadWrite permissions on the volume before allowing bidirectional mounts, as a defense in depth / validation case, although creating mounts should also require a priviliged execution environment inside the container.
1503 lines
44 KiB
Go
1503 lines
44 KiB
Go
package nomad
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"sort"
|
|
"strings"
|
|
"time"
|
|
|
|
metrics "github.com/armon/go-metrics"
|
|
log "github.com/hashicorp/go-hclog"
|
|
memdb "github.com/hashicorp/go-memdb"
|
|
multierror "github.com/hashicorp/go-multierror"
|
|
|
|
"github.com/golang/snappy"
|
|
"github.com/hashicorp/consul/lib"
|
|
"github.com/hashicorp/nomad/acl"
|
|
"github.com/hashicorp/nomad/helper"
|
|
"github.com/hashicorp/nomad/helper/uuid"
|
|
"github.com/hashicorp/nomad/nomad/state"
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
|
"github.com/hashicorp/nomad/scheduler"
|
|
)
|
|
|
|
const (
|
|
// RegisterEnforceIndexErrPrefix is the prefix to use in errors caused by
|
|
// enforcing the job modify index during registers.
|
|
RegisterEnforceIndexErrPrefix = "Enforcing job modify index"
|
|
|
|
// DispatchPayloadSizeLimit is the maximum size of the uncompressed input
|
|
// data payload.
|
|
DispatchPayloadSizeLimit = 16 * 1024
|
|
)
|
|
|
|
var (
|
|
// vaultConstraint is the implicit constraint added to jobs requesting a
|
|
// Vault token
|
|
vaultConstraint = &structs.Constraint{
|
|
LTarget: "${attr.vault.version}",
|
|
RTarget: ">= 0.6.1",
|
|
Operand: structs.ConstraintVersion,
|
|
}
|
|
|
|
// allowRescheduleTransition is the transition that allows failed
|
|
// allocations to be force rescheduled. We create a one off
|
|
// variable to avoid creating a new object for every request.
|
|
allowForceRescheduleTransition = &structs.DesiredTransition{
|
|
ForceReschedule: helper.BoolToPtr(true),
|
|
}
|
|
)
|
|
|
|
// Job endpoint is used for job interactions
|
|
type Job struct {
|
|
srv *Server
|
|
logger log.Logger
|
|
|
|
// builtin admission controllers
|
|
mutators []jobMutator
|
|
validators []jobValidator
|
|
}
|
|
|
|
// NewJobEndpoints creates a new job endpoint with builtin admission controllers
|
|
func NewJobEndpoints(s *Server) *Job {
|
|
return &Job{
|
|
srv: s,
|
|
logger: s.logger.Named("job"),
|
|
mutators: []jobMutator{
|
|
jobConnectHook{},
|
|
jobCanonicalizer{},
|
|
jobImpliedConstraints{},
|
|
},
|
|
validators: []jobValidator{
|
|
jobConnectHook{},
|
|
jobValidate{},
|
|
},
|
|
}
|
|
}
|
|
|
|
// Register is used to upsert a job for scheduling
|
|
func (j *Job) Register(args *structs.JobRegisterRequest, reply *structs.JobRegisterResponse) error {
|
|
if done, err := j.srv.forward("Job.Register", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "register"}, time.Now())
|
|
|
|
// Validate the arguments
|
|
if args.Job == nil {
|
|
return fmt.Errorf("missing job for registration")
|
|
}
|
|
|
|
// defensive check; http layer and RPC requester should ensure namespaces are set consistently
|
|
if args.RequestNamespace() != args.Job.Namespace {
|
|
return fmt.Errorf("mismatched request namespace in request: %q, %q", args.RequestNamespace(), args.Job.Namespace)
|
|
}
|
|
|
|
// Run admission controllers
|
|
job, warnings, err := j.admissionControllers(args.Job)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
args.Job = job
|
|
|
|
// Set the warning message
|
|
reply.Warnings = structs.MergeMultierrorWarnings(warnings...)
|
|
|
|
// Check job submission permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil {
|
|
if !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilitySubmitJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Validate Volume Permsissions
|
|
for _, tg := range args.Job.TaskGroups {
|
|
for _, vol := range tg.Volumes {
|
|
if vol.Type != structs.VolumeTypeHost {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// If a volume is readonly, then we allow access if the user has ReadOnly
|
|
// or ReadWrite access to the volume. Otherwise we only allow access if
|
|
// they have ReadWrite access.
|
|
if vol.ReadOnly {
|
|
if !aclObj.AllowHostVolumeOperation(vol.Source, acl.HostVolumeCapabilityMountReadOnly) &&
|
|
!aclObj.AllowHostVolumeOperation(vol.Source, acl.HostVolumeCapabilityMountReadWrite) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
} else {
|
|
if !aclObj.AllowHostVolumeOperation(vol.Source, acl.HostVolumeCapabilityMountReadWrite) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
}
|
|
}
|
|
|
|
for _, t := range tg.Tasks {
|
|
for _, vm := range t.VolumeMounts {
|
|
vol := tg.Volumes[vm.Volume]
|
|
if vm.PropagationMode == structs.VolumeMountPropagationBidirectional &&
|
|
!aclObj.AllowHostVolumeOperation(vol.Source, acl.HostVolumeCapabilityMountReadWrite) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Check if override is set and we do not have permissions
|
|
if args.PolicyOverride {
|
|
if !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilitySentinelOverride) {
|
|
j.logger.Warn("policy override attempted without permissions for job", "job", args.Job.ID)
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
j.logger.Warn("policy override set for job", "job", args.Job.ID)
|
|
}
|
|
}
|
|
|
|
// Lookup the job
|
|
snap, err := j.srv.State().Snapshot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
ws := memdb.NewWatchSet()
|
|
existingJob, err := snap.JobByID(ws, args.RequestNamespace(), args.Job.ID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// If EnforceIndex set, check it before trying to apply
|
|
if args.EnforceIndex {
|
|
jmi := args.JobModifyIndex
|
|
if existingJob != nil {
|
|
if jmi == 0 {
|
|
return fmt.Errorf("%s 0: job already exists", RegisterEnforceIndexErrPrefix)
|
|
} else if jmi != existingJob.JobModifyIndex {
|
|
return fmt.Errorf("%s %d: job exists with conflicting job modify index: %d",
|
|
RegisterEnforceIndexErrPrefix, jmi, existingJob.JobModifyIndex)
|
|
}
|
|
} else if jmi != 0 {
|
|
return fmt.Errorf("%s %d: job does not exist", RegisterEnforceIndexErrPrefix, jmi)
|
|
}
|
|
}
|
|
|
|
// Validate job transitions if its an update
|
|
if err := validateJobUpdate(existingJob, args.Job); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Ensure that the job has permissions for the requested Vault tokens
|
|
policies := args.Job.VaultPolicies()
|
|
if len(policies) != 0 {
|
|
vconf := j.srv.config.VaultConfig
|
|
if !vconf.IsEnabled() {
|
|
return fmt.Errorf("Vault not enabled and Vault policies requested")
|
|
}
|
|
|
|
// Have to check if the user has permissions
|
|
if !vconf.AllowsUnauthenticated() {
|
|
if args.Job.VaultToken == "" {
|
|
return fmt.Errorf("Vault policies requested but missing Vault Token")
|
|
}
|
|
|
|
vault := j.srv.vault
|
|
s, err := vault.LookupToken(context.Background(), args.Job.VaultToken)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
allowedPolicies, err := PoliciesFrom(s)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// If we are given a root token it can access all policies
|
|
if !lib.StrContains(allowedPolicies, "root") {
|
|
flatPolicies := structs.VaultPoliciesSet(policies)
|
|
subset, offending := helper.SliceStringIsSubset(allowedPolicies, flatPolicies)
|
|
if !subset {
|
|
return fmt.Errorf("Passed Vault Token doesn't allow access to the following policies: %s",
|
|
strings.Join(offending, ", "))
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Enforce Sentinel policies
|
|
policyWarnings, err := j.enforceSubmitJob(args.PolicyOverride, args.Job)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if policyWarnings != nil {
|
|
warnings = append(warnings, policyWarnings)
|
|
reply.Warnings = structs.MergeMultierrorWarnings(warnings...)
|
|
}
|
|
|
|
// Clear the Vault token
|
|
args.Job.VaultToken = ""
|
|
|
|
// Check if the job has changed at all
|
|
if existingJob == nil || existingJob.SpecChanged(args.Job) {
|
|
// Set the submit time
|
|
args.Job.SetSubmitTime()
|
|
|
|
// Commit this update via Raft
|
|
fsmErr, index, err := j.srv.raftApply(structs.JobRegisterRequestType, args)
|
|
if err, ok := fsmErr.(error); ok && err != nil {
|
|
j.logger.Error("registering job failed", "error", err, "fsm", true)
|
|
return err
|
|
}
|
|
if err != nil {
|
|
j.logger.Error("registering job failed", "error", err, "raft", true)
|
|
return err
|
|
}
|
|
|
|
// Populate the reply with job information
|
|
reply.JobModifyIndex = index
|
|
} else {
|
|
reply.JobModifyIndex = existingJob.JobModifyIndex
|
|
}
|
|
|
|
// If the job is periodic or parameterized, we don't create an eval.
|
|
if args.Job.IsPeriodic() || args.Job.IsParameterized() {
|
|
return nil
|
|
}
|
|
|
|
// Create a new evaluation
|
|
now := time.Now().UTC().UnixNano()
|
|
eval := &structs.Evaluation{
|
|
ID: uuid.Generate(),
|
|
Namespace: args.RequestNamespace(),
|
|
Priority: args.Job.Priority,
|
|
Type: args.Job.Type,
|
|
TriggeredBy: structs.EvalTriggerJobRegister,
|
|
JobID: args.Job.ID,
|
|
JobModifyIndex: reply.JobModifyIndex,
|
|
Status: structs.EvalStatusPending,
|
|
CreateTime: now,
|
|
ModifyTime: now,
|
|
}
|
|
update := &structs.EvalUpdateRequest{
|
|
Evals: []*structs.Evaluation{eval},
|
|
WriteRequest: structs.WriteRequest{Region: args.Region},
|
|
}
|
|
|
|
// Commit this evaluation via Raft
|
|
// XXX: There is a risk of partial failure where the JobRegister succeeds
|
|
// but that the EvalUpdate does not.
|
|
_, evalIndex, err := j.srv.raftApply(structs.EvalUpdateRequestType, update)
|
|
if err != nil {
|
|
j.logger.Error("eval create failed", "error", err, "method", "register")
|
|
return err
|
|
}
|
|
|
|
// Populate the reply with eval information
|
|
reply.EvalID = eval.ID
|
|
reply.EvalCreateIndex = evalIndex
|
|
reply.Index = evalIndex
|
|
return nil
|
|
}
|
|
|
|
// getSignalConstraint builds a suitable constraint based on the required
|
|
// signals
|
|
func getSignalConstraint(signals []string) *structs.Constraint {
|
|
sort.Strings(signals)
|
|
return &structs.Constraint{
|
|
Operand: structs.ConstraintSetContains,
|
|
LTarget: "${attr.os.signals}",
|
|
RTarget: strings.Join(signals, ","),
|
|
}
|
|
}
|
|
|
|
// Summary retrieves the summary of a job
|
|
func (j *Job) Summary(args *structs.JobSummaryRequest,
|
|
reply *structs.JobSummaryResponse) error {
|
|
|
|
if done, err := j.srv.forward("Job.Summary", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job_summary", "get_job_summary"}, time.Now())
|
|
|
|
// Check for read-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityReadJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Setup the blocking query
|
|
opts := blockingOptions{
|
|
queryOpts: &args.QueryOptions,
|
|
queryMeta: &reply.QueryMeta,
|
|
run: func(ws memdb.WatchSet, state *state.StateStore) error {
|
|
// Look for job summary
|
|
out, err := state.JobSummaryByID(ws, args.RequestNamespace(), args.JobID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Setup the output
|
|
reply.JobSummary = out
|
|
if out != nil {
|
|
reply.Index = out.ModifyIndex
|
|
} else {
|
|
// Use the last index that affected the job_summary table
|
|
index, err := state.Index("job_summary")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
reply.Index = index
|
|
}
|
|
|
|
// Set the query response
|
|
j.srv.setQueryMeta(&reply.QueryMeta)
|
|
return nil
|
|
}}
|
|
return j.srv.blockingRPC(&opts)
|
|
}
|
|
|
|
// Validate validates a job
|
|
func (j *Job) Validate(args *structs.JobValidateRequest, reply *structs.JobValidateResponse) error {
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "validate"}, time.Now())
|
|
|
|
// defensive check; http layer and RPC requester should ensure namespaces are set consistently
|
|
if args.RequestNamespace() != args.Job.Namespace {
|
|
return fmt.Errorf("mismatched request namespace in request: %q, %q", args.RequestNamespace(), args.Job.Namespace)
|
|
}
|
|
|
|
job, mutateWarnings, err := j.admissionMutators(args.Job)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
args.Job = job
|
|
|
|
// Check for read-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityReadJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Validate the job and capture any warnings
|
|
validateWarnings, err := j.admissionValidators(args.Job)
|
|
if err != nil {
|
|
if merr, ok := err.(*multierror.Error); ok {
|
|
for _, err := range merr.Errors {
|
|
reply.ValidationErrors = append(reply.ValidationErrors, err.Error())
|
|
}
|
|
reply.Error = merr.Error()
|
|
} else {
|
|
reply.ValidationErrors = append(reply.ValidationErrors, err.Error())
|
|
reply.Error = err.Error()
|
|
}
|
|
}
|
|
|
|
validateWarnings = append(validateWarnings, mutateWarnings...)
|
|
|
|
// Set the warning message
|
|
reply.Warnings = structs.MergeMultierrorWarnings(validateWarnings...)
|
|
reply.DriverConfigValidated = true
|
|
return nil
|
|
}
|
|
|
|
// Revert is used to revert the job to a prior version
|
|
func (j *Job) Revert(args *structs.JobRevertRequest, reply *structs.JobRegisterResponse) error {
|
|
if done, err := j.srv.forward("Job.Revert", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "revert"}, time.Now())
|
|
|
|
// Check for submit-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilitySubmitJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Validate the arguments
|
|
if args.JobID == "" {
|
|
return fmt.Errorf("missing job ID for revert")
|
|
}
|
|
|
|
// Lookup the job by version
|
|
snap, err := j.srv.fsm.State().Snapshot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
ws := memdb.NewWatchSet()
|
|
cur, err := snap.JobByID(ws, args.RequestNamespace(), args.JobID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if cur == nil {
|
|
return fmt.Errorf("job %q not found", args.JobID)
|
|
}
|
|
if args.JobVersion == cur.Version {
|
|
return fmt.Errorf("can't revert to current version")
|
|
}
|
|
|
|
jobV, err := snap.JobByIDAndVersion(ws, args.RequestNamespace(), args.JobID, args.JobVersion)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if jobV == nil {
|
|
return fmt.Errorf("job %q in namespace %q at version %d not found", args.JobID, args.RequestNamespace(), args.JobVersion)
|
|
}
|
|
|
|
// Build the register request
|
|
revJob := jobV.Copy()
|
|
// Use Vault Token from revert request to perform registration of reverted job.
|
|
revJob.VaultToken = args.VaultToken
|
|
reg := &structs.JobRegisterRequest{
|
|
Job: revJob,
|
|
WriteRequest: args.WriteRequest,
|
|
}
|
|
|
|
// If the request is enforcing the existing version do a check.
|
|
if args.EnforcePriorVersion != nil {
|
|
if cur.Version != *args.EnforcePriorVersion {
|
|
return fmt.Errorf("Current job has version %d; enforcing version %d", cur.Version, *args.EnforcePriorVersion)
|
|
}
|
|
|
|
reg.EnforceIndex = true
|
|
reg.JobModifyIndex = cur.JobModifyIndex
|
|
}
|
|
|
|
// Register the version.
|
|
return j.Register(reg, reply)
|
|
}
|
|
|
|
// Stable is used to mark the job version as stable
|
|
func (j *Job) Stable(args *structs.JobStabilityRequest, reply *structs.JobStabilityResponse) error {
|
|
if done, err := j.srv.forward("Job.Stable", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "stable"}, time.Now())
|
|
|
|
// Check for read-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilitySubmitJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Validate the arguments
|
|
if args.JobID == "" {
|
|
return fmt.Errorf("missing job ID for marking job as stable")
|
|
}
|
|
|
|
// Lookup the job by version
|
|
snap, err := j.srv.fsm.State().Snapshot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
ws := memdb.NewWatchSet()
|
|
jobV, err := snap.JobByIDAndVersion(ws, args.RequestNamespace(), args.JobID, args.JobVersion)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if jobV == nil {
|
|
return fmt.Errorf("job %q in namespace %q at version %d not found", args.JobID, args.RequestNamespace(), args.JobVersion)
|
|
}
|
|
|
|
// Commit this stability request via Raft
|
|
_, modifyIndex, err := j.srv.raftApply(structs.JobStabilityRequestType, args)
|
|
if err != nil {
|
|
j.logger.Error("submitting job stability request failed", "error", err)
|
|
return err
|
|
}
|
|
|
|
// Setup the reply
|
|
reply.Index = modifyIndex
|
|
return nil
|
|
}
|
|
|
|
// Evaluate is used to force a job for re-evaluation
|
|
func (j *Job) Evaluate(args *structs.JobEvaluateRequest, reply *structs.JobRegisterResponse) error {
|
|
if done, err := j.srv.forward("Job.Evaluate", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "evaluate"}, time.Now())
|
|
|
|
// Check for read-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityReadJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Validate the arguments
|
|
if args.JobID == "" {
|
|
return fmt.Errorf("missing job ID for evaluation")
|
|
}
|
|
|
|
// Lookup the job
|
|
snap, err := j.srv.fsm.State().Snapshot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
ws := memdb.NewWatchSet()
|
|
job, err := snap.JobByID(ws, args.RequestNamespace(), args.JobID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if job == nil {
|
|
return fmt.Errorf("job not found")
|
|
}
|
|
|
|
if job.IsPeriodic() {
|
|
return fmt.Errorf("can't evaluate periodic job")
|
|
} else if job.IsParameterized() {
|
|
return fmt.Errorf("can't evaluate parameterized job")
|
|
}
|
|
|
|
forceRescheduleAllocs := make(map[string]*structs.DesiredTransition)
|
|
|
|
if args.EvalOptions.ForceReschedule {
|
|
// Find any failed allocs that could be force rescheduled
|
|
allocs, err := snap.AllocsByJob(ws, args.RequestNamespace(), args.JobID, false)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
for _, alloc := range allocs {
|
|
taskGroup := job.LookupTaskGroup(alloc.TaskGroup)
|
|
// Forcing rescheduling is only allowed if task group has rescheduling enabled
|
|
if taskGroup == nil || !taskGroup.ReschedulePolicy.Enabled() {
|
|
continue
|
|
}
|
|
|
|
if alloc.NextAllocation == "" && alloc.ClientStatus == structs.AllocClientStatusFailed && !alloc.DesiredTransition.ShouldForceReschedule() {
|
|
forceRescheduleAllocs[alloc.ID] = allowForceRescheduleTransition
|
|
}
|
|
}
|
|
}
|
|
|
|
// Create a new evaluation
|
|
now := time.Now().UTC().UnixNano()
|
|
eval := &structs.Evaluation{
|
|
ID: uuid.Generate(),
|
|
Namespace: args.RequestNamespace(),
|
|
Priority: job.Priority,
|
|
Type: job.Type,
|
|
TriggeredBy: structs.EvalTriggerJobRegister,
|
|
JobID: job.ID,
|
|
JobModifyIndex: job.ModifyIndex,
|
|
Status: structs.EvalStatusPending,
|
|
CreateTime: now,
|
|
ModifyTime: now,
|
|
}
|
|
|
|
// Create a AllocUpdateDesiredTransitionRequest request with the eval and any forced rescheduled allocs
|
|
updateTransitionReq := &structs.AllocUpdateDesiredTransitionRequest{
|
|
Allocs: forceRescheduleAllocs,
|
|
Evals: []*structs.Evaluation{eval},
|
|
}
|
|
_, evalIndex, err := j.srv.raftApply(structs.AllocUpdateDesiredTransitionRequestType, updateTransitionReq)
|
|
|
|
if err != nil {
|
|
j.logger.Error("eval create failed", "error", err, "method", "evaluate")
|
|
return err
|
|
}
|
|
|
|
// Setup the reply
|
|
reply.EvalID = eval.ID
|
|
reply.EvalCreateIndex = evalIndex
|
|
reply.JobModifyIndex = job.ModifyIndex
|
|
reply.Index = evalIndex
|
|
return nil
|
|
}
|
|
|
|
// Deregister is used to remove a job the cluster.
|
|
func (j *Job) Deregister(args *structs.JobDeregisterRequest, reply *structs.JobDeregisterResponse) error {
|
|
if done, err := j.srv.forward("Job.Deregister", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "deregister"}, time.Now())
|
|
|
|
// Check for submit-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilitySubmitJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Validate the arguments
|
|
if args.JobID == "" {
|
|
return fmt.Errorf("missing job ID for deregistering")
|
|
}
|
|
|
|
// Lookup the job
|
|
snap, err := j.srv.fsm.State().Snapshot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
ws := memdb.NewWatchSet()
|
|
job, err := snap.JobByID(ws, args.RequestNamespace(), args.JobID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Commit this update via Raft
|
|
_, index, err := j.srv.raftApply(structs.JobDeregisterRequestType, args)
|
|
if err != nil {
|
|
j.logger.Error("deregister failed", "error", err)
|
|
return err
|
|
}
|
|
|
|
// Populate the reply with job information
|
|
reply.JobModifyIndex = index
|
|
|
|
// If the job is periodic or parameterized, we don't create an eval.
|
|
if job != nil && (job.IsPeriodic() || job.IsParameterized()) {
|
|
return nil
|
|
}
|
|
|
|
// Create a new evaluation
|
|
// XXX: The job priority / type is strange for this, since it's not a high
|
|
// priority even if the job was.
|
|
now := time.Now().UTC().UnixNano()
|
|
eval := &structs.Evaluation{
|
|
ID: uuid.Generate(),
|
|
Namespace: args.RequestNamespace(),
|
|
Priority: structs.JobDefaultPriority,
|
|
Type: structs.JobTypeService,
|
|
TriggeredBy: structs.EvalTriggerJobDeregister,
|
|
JobID: args.JobID,
|
|
JobModifyIndex: index,
|
|
Status: structs.EvalStatusPending,
|
|
CreateTime: now,
|
|
ModifyTime: now,
|
|
}
|
|
update := &structs.EvalUpdateRequest{
|
|
Evals: []*structs.Evaluation{eval},
|
|
WriteRequest: structs.WriteRequest{Region: args.Region},
|
|
}
|
|
|
|
// Commit this evaluation via Raft
|
|
_, evalIndex, err := j.srv.raftApply(structs.EvalUpdateRequestType, update)
|
|
if err != nil {
|
|
j.logger.Error("eval create failed", "error", err, "method", "deregister")
|
|
return err
|
|
}
|
|
|
|
// Populate the reply with eval information
|
|
reply.EvalID = eval.ID
|
|
reply.EvalCreateIndex = evalIndex
|
|
reply.Index = evalIndex
|
|
return nil
|
|
}
|
|
|
|
// BatchDeregister is used to remove a set of jobs from the cluster.
|
|
func (j *Job) BatchDeregister(args *structs.JobBatchDeregisterRequest, reply *structs.JobBatchDeregisterResponse) error {
|
|
if done, err := j.srv.forward("Job.BatchDeregister", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "batch_deregister"}, time.Now())
|
|
|
|
// Resolve the ACL token
|
|
aclObj, err := j.srv.ResolveToken(args.AuthToken)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Validate the arguments
|
|
if len(args.Jobs) == 0 {
|
|
return fmt.Errorf("given no jobs to deregister")
|
|
}
|
|
if len(args.Evals) != 0 {
|
|
return fmt.Errorf("evaluations should not be populated")
|
|
}
|
|
|
|
// Loop through checking for permissions
|
|
for jobNS := range args.Jobs {
|
|
// Check for submit-job permissions
|
|
if aclObj != nil && !aclObj.AllowNsOp(jobNS.Namespace, acl.NamespaceCapabilitySubmitJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
}
|
|
|
|
// Grab a snapshot
|
|
snap, err := j.srv.fsm.State().Snapshot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Loop through to create evals
|
|
for jobNS, options := range args.Jobs {
|
|
if options == nil {
|
|
return fmt.Errorf("no deregister options provided for %v", jobNS)
|
|
}
|
|
|
|
job, err := snap.JobByID(nil, jobNS.Namespace, jobNS.ID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// If the job is periodic or parameterized, we don't create an eval.
|
|
if job != nil && (job.IsPeriodic() || job.IsParameterized()) {
|
|
continue
|
|
}
|
|
|
|
priority := structs.JobDefaultPriority
|
|
jtype := structs.JobTypeService
|
|
if job != nil {
|
|
priority = job.Priority
|
|
jtype = job.Type
|
|
}
|
|
|
|
// Create a new evaluation
|
|
now := time.Now().UTC().UnixNano()
|
|
eval := &structs.Evaluation{
|
|
ID: uuid.Generate(),
|
|
Namespace: jobNS.Namespace,
|
|
Priority: priority,
|
|
Type: jtype,
|
|
TriggeredBy: structs.EvalTriggerJobDeregister,
|
|
JobID: jobNS.ID,
|
|
Status: structs.EvalStatusPending,
|
|
CreateTime: now,
|
|
ModifyTime: now,
|
|
}
|
|
args.Evals = append(args.Evals, eval)
|
|
}
|
|
|
|
// Commit this update via Raft
|
|
_, index, err := j.srv.raftApply(structs.JobBatchDeregisterRequestType, args)
|
|
if err != nil {
|
|
j.logger.Error("batch deregister failed", "error", err)
|
|
return err
|
|
}
|
|
|
|
reply.Index = index
|
|
return nil
|
|
}
|
|
|
|
// GetJob is used to request information about a specific job
|
|
func (j *Job) GetJob(args *structs.JobSpecificRequest,
|
|
reply *structs.SingleJobResponse) error {
|
|
if done, err := j.srv.forward("Job.GetJob", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "get_job"}, time.Now())
|
|
|
|
// Check for read-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityReadJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Setup the blocking query
|
|
opts := blockingOptions{
|
|
queryOpts: &args.QueryOptions,
|
|
queryMeta: &reply.QueryMeta,
|
|
run: func(ws memdb.WatchSet, state *state.StateStore) error {
|
|
// Look for the job
|
|
out, err := state.JobByID(ws, args.RequestNamespace(), args.JobID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Setup the output
|
|
reply.Job = out
|
|
if out != nil {
|
|
reply.Index = out.ModifyIndex
|
|
} else {
|
|
// Use the last index that affected the nodes table
|
|
index, err := state.Index("jobs")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
reply.Index = index
|
|
}
|
|
|
|
// Set the query response
|
|
j.srv.setQueryMeta(&reply.QueryMeta)
|
|
return nil
|
|
}}
|
|
return j.srv.blockingRPC(&opts)
|
|
}
|
|
|
|
// GetJobVersions is used to retrieve all tracked versions of a job.
|
|
func (j *Job) GetJobVersions(args *structs.JobVersionsRequest,
|
|
reply *structs.JobVersionsResponse) error {
|
|
if done, err := j.srv.forward("Job.GetJobVersions", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "get_job_versions"}, time.Now())
|
|
|
|
// Check for read-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityReadJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Setup the blocking query
|
|
opts := blockingOptions{
|
|
queryOpts: &args.QueryOptions,
|
|
queryMeta: &reply.QueryMeta,
|
|
run: func(ws memdb.WatchSet, state *state.StateStore) error {
|
|
// Look for the job
|
|
out, err := state.JobVersionsByID(ws, args.RequestNamespace(), args.JobID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Setup the output
|
|
reply.Versions = out
|
|
if len(out) != 0 {
|
|
reply.Index = out[0].ModifyIndex
|
|
|
|
// Compute the diffs
|
|
if args.Diffs {
|
|
for i := 0; i < len(out)-1; i++ {
|
|
old, new := out[i+1], out[i]
|
|
d, err := old.Diff(new, true)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to create job diff: %v", err)
|
|
}
|
|
reply.Diffs = append(reply.Diffs, d)
|
|
}
|
|
}
|
|
} else {
|
|
// Use the last index that affected the nodes table
|
|
index, err := state.Index("job_version")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
reply.Index = index
|
|
}
|
|
|
|
// Set the query response
|
|
j.srv.setQueryMeta(&reply.QueryMeta)
|
|
return nil
|
|
}}
|
|
return j.srv.blockingRPC(&opts)
|
|
}
|
|
|
|
// List is used to list the jobs registered in the system
|
|
func (j *Job) List(args *structs.JobListRequest,
|
|
reply *structs.JobListResponse) error {
|
|
if done, err := j.srv.forward("Job.List", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "list"}, time.Now())
|
|
|
|
// Check for list-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityListJobs) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Setup the blocking query
|
|
opts := blockingOptions{
|
|
queryOpts: &args.QueryOptions,
|
|
queryMeta: &reply.QueryMeta,
|
|
run: func(ws memdb.WatchSet, state *state.StateStore) error {
|
|
// Capture all the jobs
|
|
var err error
|
|
var iter memdb.ResultIterator
|
|
if prefix := args.QueryOptions.Prefix; prefix != "" {
|
|
iter, err = state.JobsByIDPrefix(ws, args.RequestNamespace(), prefix)
|
|
} else {
|
|
iter, err = state.JobsByNamespace(ws, args.RequestNamespace())
|
|
}
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
var jobs []*structs.JobListStub
|
|
for {
|
|
raw := iter.Next()
|
|
if raw == nil {
|
|
break
|
|
}
|
|
job := raw.(*structs.Job)
|
|
summary, err := state.JobSummaryByID(ws, args.RequestNamespace(), job.ID)
|
|
if err != nil {
|
|
return fmt.Errorf("unable to look up summary for job: %v", job.ID)
|
|
}
|
|
jobs = append(jobs, job.Stub(summary))
|
|
}
|
|
reply.Jobs = jobs
|
|
|
|
// Use the last index that affected the jobs table or summary
|
|
jindex, err := state.Index("jobs")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
sindex, err := state.Index("job_summary")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
reply.Index = helper.Uint64Max(jindex, sindex)
|
|
|
|
// Set the query response
|
|
j.srv.setQueryMeta(&reply.QueryMeta)
|
|
return nil
|
|
}}
|
|
return j.srv.blockingRPC(&opts)
|
|
}
|
|
|
|
// Allocations is used to list the allocations for a job
|
|
func (j *Job) Allocations(args *structs.JobSpecificRequest,
|
|
reply *structs.JobAllocationsResponse) error {
|
|
if done, err := j.srv.forward("Job.Allocations", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "allocations"}, time.Now())
|
|
|
|
// Check for read-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityReadJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Setup the blocking query
|
|
opts := blockingOptions{
|
|
queryOpts: &args.QueryOptions,
|
|
queryMeta: &reply.QueryMeta,
|
|
run: func(ws memdb.WatchSet, state *state.StateStore) error {
|
|
// Capture the allocations
|
|
allocs, err := state.AllocsByJob(ws, args.RequestNamespace(), args.JobID, args.All)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Convert to stubs
|
|
if len(allocs) > 0 {
|
|
reply.Allocations = make([]*structs.AllocListStub, 0, len(allocs))
|
|
for _, alloc := range allocs {
|
|
reply.Allocations = append(reply.Allocations, alloc.Stub())
|
|
}
|
|
}
|
|
|
|
// Use the last index that affected the allocs table
|
|
index, err := state.Index("allocs")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
reply.Index = index
|
|
|
|
// Set the query response
|
|
j.srv.setQueryMeta(&reply.QueryMeta)
|
|
return nil
|
|
|
|
}}
|
|
return j.srv.blockingRPC(&opts)
|
|
}
|
|
|
|
// Evaluations is used to list the evaluations for a job
|
|
func (j *Job) Evaluations(args *structs.JobSpecificRequest,
|
|
reply *structs.JobEvaluationsResponse) error {
|
|
if done, err := j.srv.forward("Job.Evaluations", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "evaluations"}, time.Now())
|
|
|
|
// Check for read-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityReadJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Setup the blocking query
|
|
opts := blockingOptions{
|
|
queryOpts: &args.QueryOptions,
|
|
queryMeta: &reply.QueryMeta,
|
|
run: func(ws memdb.WatchSet, state *state.StateStore) error {
|
|
// Capture the evals
|
|
var err error
|
|
reply.Evaluations, err = state.EvalsByJob(ws, args.RequestNamespace(), args.JobID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Use the last index that affected the evals table
|
|
index, err := state.Index("evals")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
reply.Index = index
|
|
|
|
// Set the query response
|
|
j.srv.setQueryMeta(&reply.QueryMeta)
|
|
return nil
|
|
}}
|
|
|
|
return j.srv.blockingRPC(&opts)
|
|
}
|
|
|
|
// Deployments is used to list the deployments for a job
|
|
func (j *Job) Deployments(args *structs.JobSpecificRequest,
|
|
reply *structs.DeploymentListResponse) error {
|
|
if done, err := j.srv.forward("Job.Deployments", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "deployments"}, time.Now())
|
|
|
|
// Check for read-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityReadJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Setup the blocking query
|
|
opts := blockingOptions{
|
|
queryOpts: &args.QueryOptions,
|
|
queryMeta: &reply.QueryMeta,
|
|
run: func(ws memdb.WatchSet, state *state.StateStore) error {
|
|
// Capture the deployments
|
|
deploys, err := state.DeploymentsByJobID(ws, args.RequestNamespace(), args.JobID, args.All)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Use the last index that affected the deployment table
|
|
index, err := state.Index("deployment")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
reply.Index = index
|
|
reply.Deployments = deploys
|
|
|
|
// Set the query response
|
|
j.srv.setQueryMeta(&reply.QueryMeta)
|
|
return nil
|
|
|
|
}}
|
|
return j.srv.blockingRPC(&opts)
|
|
}
|
|
|
|
// LatestDeployment is used to retrieve the latest deployment for a job
|
|
func (j *Job) LatestDeployment(args *structs.JobSpecificRequest,
|
|
reply *structs.SingleDeploymentResponse) error {
|
|
if done, err := j.srv.forward("Job.LatestDeployment", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "latest_deployment"}, time.Now())
|
|
|
|
// Check for read-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityReadJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Setup the blocking query
|
|
opts := blockingOptions{
|
|
queryOpts: &args.QueryOptions,
|
|
queryMeta: &reply.QueryMeta,
|
|
run: func(ws memdb.WatchSet, state *state.StateStore) error {
|
|
// Capture the deployments
|
|
deploys, err := state.DeploymentsByJobID(ws, args.RequestNamespace(), args.JobID, args.All)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Use the last index that affected the deployment table
|
|
index, err := state.Index("deployment")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
reply.Index = index
|
|
if len(deploys) > 0 {
|
|
sort.Slice(deploys, func(i, j int) bool {
|
|
return deploys[i].CreateIndex > deploys[j].CreateIndex
|
|
})
|
|
reply.Deployment = deploys[0]
|
|
}
|
|
|
|
// Set the query response
|
|
j.srv.setQueryMeta(&reply.QueryMeta)
|
|
return nil
|
|
|
|
}}
|
|
return j.srv.blockingRPC(&opts)
|
|
}
|
|
|
|
// Plan is used to cause a dry-run evaluation of the Job and return the results
|
|
// with a potential diff containing annotations.
|
|
func (j *Job) Plan(args *structs.JobPlanRequest, reply *structs.JobPlanResponse) error {
|
|
if done, err := j.srv.forward("Job.Plan", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "plan"}, time.Now())
|
|
|
|
// Validate the arguments
|
|
if args.Job == nil {
|
|
return fmt.Errorf("Job required for plan")
|
|
}
|
|
|
|
// Run admission controllers
|
|
job, warnings, err := j.admissionControllers(args.Job)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
args.Job = job
|
|
|
|
// Set the warning message
|
|
reply.Warnings = structs.MergeMultierrorWarnings(warnings...)
|
|
|
|
// Check job submission permissions, which we assume is the same for plan
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil {
|
|
if !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilitySubmitJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
// Check if override is set and we do not have permissions
|
|
if args.PolicyOverride {
|
|
if !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilitySentinelOverride) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
}
|
|
}
|
|
|
|
// Enforce Sentinel policies
|
|
policyWarnings, err := j.enforceSubmitJob(args.PolicyOverride, args.Job)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if policyWarnings != nil {
|
|
warnings = append(warnings, policyWarnings)
|
|
reply.Warnings = structs.MergeMultierrorWarnings(warnings...)
|
|
}
|
|
|
|
// Acquire a snapshot of the state
|
|
snap, err := j.srv.fsm.State().Snapshot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Get the original job
|
|
ws := memdb.NewWatchSet()
|
|
oldJob, err := snap.JobByID(ws, args.RequestNamespace(), args.Job.ID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
var index uint64
|
|
var updatedIndex uint64
|
|
|
|
if oldJob != nil {
|
|
index = oldJob.JobModifyIndex
|
|
|
|
// We want to reuse deployments where possible, so only insert the job if
|
|
// it has changed or the job didn't exist
|
|
if oldJob.SpecChanged(args.Job) {
|
|
// Insert the updated Job into the snapshot
|
|
updatedIndex = oldJob.JobModifyIndex + 1
|
|
snap.UpsertJob(updatedIndex, args.Job)
|
|
}
|
|
} else if oldJob == nil {
|
|
// Insert the updated Job into the snapshot
|
|
snap.UpsertJob(100, args.Job)
|
|
}
|
|
|
|
// Create an eval and mark it as requiring annotations and insert that as well
|
|
now := time.Now().UTC().UnixNano()
|
|
eval := &structs.Evaluation{
|
|
ID: uuid.Generate(),
|
|
Namespace: args.RequestNamespace(),
|
|
Priority: args.Job.Priority,
|
|
Type: args.Job.Type,
|
|
TriggeredBy: structs.EvalTriggerJobRegister,
|
|
JobID: args.Job.ID,
|
|
JobModifyIndex: updatedIndex,
|
|
Status: structs.EvalStatusPending,
|
|
AnnotatePlan: true,
|
|
// Timestamps are added for consistency but this eval is never persisted
|
|
CreateTime: now,
|
|
ModifyTime: now,
|
|
}
|
|
|
|
snap.UpsertEvals(100, []*structs.Evaluation{eval})
|
|
|
|
// Create an in-memory Planner that returns no errors and stores the
|
|
// submitted plan and created evals.
|
|
planner := &scheduler.Harness{
|
|
State: &snap.StateStore,
|
|
}
|
|
|
|
// Create the scheduler and run it
|
|
sched, err := scheduler.NewScheduler(eval.Type, j.logger, snap, planner)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := sched.Process(eval); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Annotate and store the diff
|
|
if plans := len(planner.Plans); plans != 1 {
|
|
return fmt.Errorf("scheduler resulted in an unexpected number of plans: %v", plans)
|
|
}
|
|
annotations := planner.Plans[0].Annotations
|
|
if args.Diff {
|
|
jobDiff, err := oldJob.Diff(args.Job, true)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to create job diff: %v", err)
|
|
}
|
|
|
|
if err := scheduler.Annotate(jobDiff, annotations); err != nil {
|
|
return fmt.Errorf("failed to annotate job diff: %v", err)
|
|
}
|
|
reply.Diff = jobDiff
|
|
}
|
|
|
|
// Grab the failures
|
|
if len(planner.Evals) != 1 {
|
|
return fmt.Errorf("scheduler resulted in an unexpected number of eval updates: %v", planner.Evals)
|
|
}
|
|
updatedEval := planner.Evals[0]
|
|
|
|
// If it is a periodic job calculate the next launch
|
|
if args.Job.IsPeriodic() && args.Job.Periodic.Enabled {
|
|
reply.NextPeriodicLaunch, err = args.Job.Periodic.Next(time.Now().In(args.Job.Periodic.GetLocation()))
|
|
if err != nil {
|
|
return fmt.Errorf("Failed to parse cron expression: %v", err)
|
|
}
|
|
}
|
|
|
|
reply.FailedTGAllocs = updatedEval.FailedTGAllocs
|
|
reply.JobModifyIndex = index
|
|
reply.Annotations = annotations
|
|
reply.CreatedEvals = planner.CreateEvals
|
|
reply.Index = index
|
|
return nil
|
|
}
|
|
|
|
// validateJobUpdate ensures updates to a job are valid.
|
|
func validateJobUpdate(old, new *structs.Job) error {
|
|
// Validate Dispatch not set on new Jobs
|
|
if old == nil {
|
|
if new.Dispatched {
|
|
return fmt.Errorf("job can't be submitted with 'Dispatched' set")
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Type transitions are disallowed
|
|
if old.Type != new.Type {
|
|
return fmt.Errorf("cannot update job from type %q to %q", old.Type, new.Type)
|
|
}
|
|
|
|
// Transitioning to/from periodic is disallowed
|
|
if old.IsPeriodic() && !new.IsPeriodic() {
|
|
return fmt.Errorf("cannot update periodic job to being non-periodic")
|
|
}
|
|
if new.IsPeriodic() && !old.IsPeriodic() {
|
|
return fmt.Errorf("cannot update non-periodic job to being periodic")
|
|
}
|
|
|
|
// Transitioning to/from parameterized is disallowed
|
|
if old.IsParameterized() && !new.IsParameterized() {
|
|
return fmt.Errorf("cannot update non-parameterized job to being parameterized")
|
|
}
|
|
if new.IsParameterized() && !old.IsParameterized() {
|
|
return fmt.Errorf("cannot update parameterized job to being non-parameterized")
|
|
}
|
|
|
|
if old.Dispatched != new.Dispatched {
|
|
return fmt.Errorf("field 'Dispatched' is read-only")
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// Dispatch a parameterized job.
|
|
func (j *Job) Dispatch(args *structs.JobDispatchRequest, reply *structs.JobDispatchResponse) error {
|
|
if done, err := j.srv.forward("Job.Dispatch", args, args, reply); done {
|
|
return err
|
|
}
|
|
defer metrics.MeasureSince([]string{"nomad", "job", "dispatch"}, time.Now())
|
|
|
|
// Check for submit-job permissions
|
|
if aclObj, err := j.srv.ResolveToken(args.AuthToken); err != nil {
|
|
return err
|
|
} else if aclObj != nil && !aclObj.AllowNsOp(args.RequestNamespace(), acl.NamespaceCapabilityDispatchJob) {
|
|
return structs.ErrPermissionDenied
|
|
}
|
|
|
|
// Lookup the parameterized job
|
|
if args.JobID == "" {
|
|
return fmt.Errorf("missing parameterized job ID")
|
|
}
|
|
|
|
snap, err := j.srv.fsm.State().Snapshot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
ws := memdb.NewWatchSet()
|
|
parameterizedJob, err := snap.JobByID(ws, args.RequestNamespace(), args.JobID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if parameterizedJob == nil {
|
|
return fmt.Errorf("parameterized job not found")
|
|
}
|
|
|
|
if !parameterizedJob.IsParameterized() {
|
|
return fmt.Errorf("Specified job %q is not a parameterized job", args.JobID)
|
|
}
|
|
|
|
if parameterizedJob.Stop {
|
|
return fmt.Errorf("Specified job %q is stopped", args.JobID)
|
|
}
|
|
|
|
// Validate the arguments
|
|
if err := validateDispatchRequest(args, parameterizedJob); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Derive the child job and commit it via Raft
|
|
dispatchJob := parameterizedJob.Copy()
|
|
dispatchJob.ID = structs.DispatchedID(parameterizedJob.ID, time.Now())
|
|
dispatchJob.ParentID = parameterizedJob.ID
|
|
dispatchJob.Name = dispatchJob.ID
|
|
dispatchJob.SetSubmitTime()
|
|
dispatchJob.Dispatched = true
|
|
|
|
// Merge in the meta data
|
|
for k, v := range args.Meta {
|
|
if dispatchJob.Meta == nil {
|
|
dispatchJob.Meta = make(map[string]string, len(args.Meta))
|
|
}
|
|
dispatchJob.Meta[k] = v
|
|
}
|
|
|
|
// Compress the payload
|
|
dispatchJob.Payload = snappy.Encode(nil, args.Payload)
|
|
|
|
regReq := &structs.JobRegisterRequest{
|
|
Job: dispatchJob,
|
|
WriteRequest: args.WriteRequest,
|
|
}
|
|
|
|
// Commit this update via Raft
|
|
fsmErr, jobCreateIndex, err := j.srv.raftApply(structs.JobRegisterRequestType, regReq)
|
|
if err, ok := fsmErr.(error); ok && err != nil {
|
|
j.logger.Error("dispatched job register failed", "error", err, "fsm", true)
|
|
return err
|
|
}
|
|
if err != nil {
|
|
j.logger.Error("dispatched job register failed", "error", err, "raft", true)
|
|
return err
|
|
}
|
|
|
|
reply.JobCreateIndex = jobCreateIndex
|
|
reply.DispatchedJobID = dispatchJob.ID
|
|
reply.Index = jobCreateIndex
|
|
|
|
// If the job is periodic, we don't create an eval.
|
|
if !dispatchJob.IsPeriodic() {
|
|
// Create a new evaluation
|
|
now := time.Now().UTC().UnixNano()
|
|
eval := &structs.Evaluation{
|
|
ID: uuid.Generate(),
|
|
Namespace: args.RequestNamespace(),
|
|
Priority: dispatchJob.Priority,
|
|
Type: dispatchJob.Type,
|
|
TriggeredBy: structs.EvalTriggerJobRegister,
|
|
JobID: dispatchJob.ID,
|
|
JobModifyIndex: jobCreateIndex,
|
|
Status: structs.EvalStatusPending,
|
|
CreateTime: now,
|
|
ModifyTime: now,
|
|
}
|
|
update := &structs.EvalUpdateRequest{
|
|
Evals: []*structs.Evaluation{eval},
|
|
WriteRequest: structs.WriteRequest{Region: args.Region},
|
|
}
|
|
|
|
// Commit this evaluation via Raft
|
|
_, evalIndex, err := j.srv.raftApply(structs.EvalUpdateRequestType, update)
|
|
if err != nil {
|
|
j.logger.Error("eval create failed", "error", err, "method", "dispatch")
|
|
return err
|
|
}
|
|
|
|
// Setup the reply
|
|
reply.EvalID = eval.ID
|
|
reply.EvalCreateIndex = evalIndex
|
|
reply.Index = evalIndex
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// validateDispatchRequest returns whether the request is valid given the
|
|
// parameterized job.
|
|
func validateDispatchRequest(req *structs.JobDispatchRequest, job *structs.Job) error {
|
|
// Check the payload constraint is met
|
|
hasInputData := len(req.Payload) != 0
|
|
if job.ParameterizedJob.Payload == structs.DispatchPayloadRequired && !hasInputData {
|
|
return fmt.Errorf("Payload is not provided but required by parameterized job")
|
|
} else if job.ParameterizedJob.Payload == structs.DispatchPayloadForbidden && hasInputData {
|
|
return fmt.Errorf("Payload provided but forbidden by parameterized job")
|
|
}
|
|
|
|
// Check the payload doesn't exceed the size limit
|
|
if l := len(req.Payload); l > DispatchPayloadSizeLimit {
|
|
return fmt.Errorf("Payload exceeds maximum size; %d > %d", l, DispatchPayloadSizeLimit)
|
|
}
|
|
|
|
// Check if the metadata is a set
|
|
keys := make(map[string]struct{}, len(req.Meta))
|
|
for k := range keys {
|
|
if _, ok := keys[k]; ok {
|
|
return fmt.Errorf("Duplicate key %q in passed metadata", k)
|
|
}
|
|
keys[k] = struct{}{}
|
|
}
|
|
|
|
required := helper.SliceStringToSet(job.ParameterizedJob.MetaRequired)
|
|
optional := helper.SliceStringToSet(job.ParameterizedJob.MetaOptional)
|
|
|
|
// Check the metadata key constraints are met
|
|
unpermitted := make(map[string]struct{})
|
|
for k := range req.Meta {
|
|
_, req := required[k]
|
|
_, opt := optional[k]
|
|
if !req && !opt {
|
|
unpermitted[k] = struct{}{}
|
|
}
|
|
}
|
|
|
|
if len(unpermitted) != 0 {
|
|
flat := make([]string, 0, len(unpermitted))
|
|
for k := range unpermitted {
|
|
flat = append(flat, k)
|
|
}
|
|
|
|
return fmt.Errorf("Dispatch request included unpermitted metadata keys: %v", flat)
|
|
}
|
|
|
|
missing := make(map[string]struct{})
|
|
for _, k := range job.ParameterizedJob.MetaRequired {
|
|
if _, ok := req.Meta[k]; !ok {
|
|
missing[k] = struct{}{}
|
|
}
|
|
}
|
|
|
|
if len(missing) != 0 {
|
|
flat := make([]string, 0, len(missing))
|
|
for k := range missing {
|
|
flat = append(flat, k)
|
|
}
|
|
|
|
return fmt.Errorf("Dispatch did not provide required meta keys: %v", flat)
|
|
}
|
|
|
|
return nil
|
|
}
|